var/home/core/zuul-output/0000755000175000017500000000000015114343116014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114370515015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006710323715114370507017711 0ustar rootrootDec 04 17:38:57 crc systemd[1]: Starting Kubernetes Kubelet... Dec 04 17:38:57 crc restorecon[4732]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 17:38:57 crc restorecon[4732]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 04 17:38:58 crc kubenswrapper[4733]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 17:38:58 crc kubenswrapper[4733]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 04 17:38:58 crc kubenswrapper[4733]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 17:38:58 crc kubenswrapper[4733]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 17:38:58 crc kubenswrapper[4733]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 04 17:38:58 crc kubenswrapper[4733]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.180466 4733 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182879 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182893 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182898 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182901 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182905 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182908 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182912 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182916 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182921 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182926 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182930 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182933 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182938 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182949 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182953 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182957 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182960 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182964 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182968 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182971 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182975 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182978 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182982 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182986 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182990 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.182994 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183000 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183004 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183008 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183011 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183015 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183019 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183022 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183026 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183029 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183033 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183036 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183039 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183043 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183046 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183050 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183053 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183056 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183060 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183063 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183068 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183072 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183076 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183080 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183084 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183088 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183091 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183095 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183098 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183102 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183106 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183109 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183113 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183116 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183120 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183123 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183127 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183133 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183137 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183141 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183144 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183148 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183151 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183155 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183158 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.183162 4733 feature_gate.go:330] unrecognized feature gate: Example Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183643 4733 flags.go:64] FLAG: --address="0.0.0.0" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183655 4733 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183663 4733 flags.go:64] FLAG: --anonymous-auth="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183669 4733 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183674 4733 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183678 4733 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183684 4733 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183689 4733 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183693 4733 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183697 4733 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183702 4733 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183706 4733 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183710 4733 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183714 4733 flags.go:64] FLAG: --cgroup-root="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183718 4733 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183722 4733 flags.go:64] FLAG: --client-ca-file="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183726 4733 flags.go:64] FLAG: --cloud-config="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183730 4733 flags.go:64] FLAG: --cloud-provider="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183734 4733 flags.go:64] FLAG: --cluster-dns="[]" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183739 4733 flags.go:64] FLAG: --cluster-domain="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183743 4733 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183747 4733 flags.go:64] FLAG: --config-dir="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183750 4733 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183755 4733 flags.go:64] FLAG: --container-log-max-files="5" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183760 4733 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183764 4733 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183769 4733 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183773 4733 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183777 4733 flags.go:64] FLAG: --contention-profiling="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183781 4733 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183786 4733 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183803 4733 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183807 4733 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183812 4733 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183816 4733 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183820 4733 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183824 4733 flags.go:64] FLAG: --enable-load-reader="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183828 4733 flags.go:64] FLAG: --enable-server="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183833 4733 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183839 4733 flags.go:64] FLAG: --event-burst="100" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183843 4733 flags.go:64] FLAG: --event-qps="50" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183847 4733 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183851 4733 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183855 4733 flags.go:64] FLAG: --eviction-hard="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183860 4733 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183864 4733 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183868 4733 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183872 4733 flags.go:64] FLAG: --eviction-soft="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183876 4733 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183880 4733 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183884 4733 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183888 4733 flags.go:64] FLAG: --experimental-mounter-path="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183892 4733 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183896 4733 flags.go:64] FLAG: --fail-swap-on="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183900 4733 flags.go:64] FLAG: --feature-gates="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183905 4733 flags.go:64] FLAG: --file-check-frequency="20s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183909 4733 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183913 4733 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183917 4733 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183921 4733 flags.go:64] FLAG: --healthz-port="10248" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183925 4733 flags.go:64] FLAG: --help="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183930 4733 flags.go:64] FLAG: --hostname-override="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183934 4733 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183939 4733 flags.go:64] FLAG: --http-check-frequency="20s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183943 4733 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183947 4733 flags.go:64] FLAG: --image-credential-provider-config="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183951 4733 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183955 4733 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183960 4733 flags.go:64] FLAG: --image-service-endpoint="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183963 4733 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183968 4733 flags.go:64] FLAG: --kube-api-burst="100" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183972 4733 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183977 4733 flags.go:64] FLAG: --kube-api-qps="50" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183981 4733 flags.go:64] FLAG: --kube-reserved="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183985 4733 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183989 4733 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183993 4733 flags.go:64] FLAG: --kubelet-cgroups="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.183997 4733 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184001 4733 flags.go:64] FLAG: --lock-file="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184007 4733 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184012 4733 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184016 4733 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184022 4733 flags.go:64] FLAG: --log-json-split-stream="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184027 4733 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184032 4733 flags.go:64] FLAG: --log-text-split-stream="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184037 4733 flags.go:64] FLAG: --logging-format="text" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184041 4733 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184045 4733 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184049 4733 flags.go:64] FLAG: --manifest-url="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184053 4733 flags.go:64] FLAG: --manifest-url-header="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184059 4733 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184063 4733 flags.go:64] FLAG: --max-open-files="1000000" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184068 4733 flags.go:64] FLAG: --max-pods="110" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184072 4733 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184076 4733 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184080 4733 flags.go:64] FLAG: --memory-manager-policy="None" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184084 4733 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184089 4733 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184093 4733 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184098 4733 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184106 4733 flags.go:64] FLAG: --node-status-max-images="50" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184110 4733 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184115 4733 flags.go:64] FLAG: --oom-score-adj="-999" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184120 4733 flags.go:64] FLAG: --pod-cidr="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184124 4733 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184130 4733 flags.go:64] FLAG: --pod-manifest-path="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184134 4733 flags.go:64] FLAG: --pod-max-pids="-1" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184138 4733 flags.go:64] FLAG: --pods-per-core="0" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184142 4733 flags.go:64] FLAG: --port="10250" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184146 4733 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184150 4733 flags.go:64] FLAG: --provider-id="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184154 4733 flags.go:64] FLAG: --qos-reserved="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184158 4733 flags.go:64] FLAG: --read-only-port="10255" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184162 4733 flags.go:64] FLAG: --register-node="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184167 4733 flags.go:64] FLAG: --register-schedulable="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184171 4733 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184178 4733 flags.go:64] FLAG: --registry-burst="10" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184182 4733 flags.go:64] FLAG: --registry-qps="5" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184187 4733 flags.go:64] FLAG: --reserved-cpus="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184191 4733 flags.go:64] FLAG: --reserved-memory="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184196 4733 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184200 4733 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184205 4733 flags.go:64] FLAG: --rotate-certificates="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184209 4733 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184213 4733 flags.go:64] FLAG: --runonce="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184217 4733 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184222 4733 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184227 4733 flags.go:64] FLAG: --seccomp-default="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184231 4733 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184235 4733 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184239 4733 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184243 4733 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184247 4733 flags.go:64] FLAG: --storage-driver-password="root" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184251 4733 flags.go:64] FLAG: --storage-driver-secure="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184256 4733 flags.go:64] FLAG: --storage-driver-table="stats" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184269 4733 flags.go:64] FLAG: --storage-driver-user="root" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184273 4733 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184277 4733 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184282 4733 flags.go:64] FLAG: --system-cgroups="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184286 4733 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184293 4733 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184296 4733 flags.go:64] FLAG: --tls-cert-file="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184300 4733 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184305 4733 flags.go:64] FLAG: --tls-min-version="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184310 4733 flags.go:64] FLAG: --tls-private-key-file="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184314 4733 flags.go:64] FLAG: --topology-manager-policy="none" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184318 4733 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184322 4733 flags.go:64] FLAG: --topology-manager-scope="container" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184327 4733 flags.go:64] FLAG: --v="2" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184337 4733 flags.go:64] FLAG: --version="false" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184342 4733 flags.go:64] FLAG: --vmodule="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184347 4733 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184351 4733 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184488 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184494 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184499 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184503 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184507 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184511 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184515 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184518 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184522 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184525 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184529 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184532 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184536 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184540 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184545 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184549 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184553 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184557 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184562 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184567 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184571 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184575 4733 feature_gate.go:330] unrecognized feature gate: Example Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184579 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184583 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184587 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184590 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184594 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184598 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184602 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184606 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184609 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184613 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184616 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184620 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184623 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184627 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184630 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184634 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184637 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184642 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184645 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184649 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184652 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184656 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184660 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184663 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184669 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184673 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184676 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184681 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184685 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184689 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184693 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184697 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184702 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184705 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184709 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184712 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184716 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184720 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184724 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184728 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184731 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184735 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184738 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184742 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184745 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184749 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184752 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184756 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.184759 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.184766 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.195824 4733 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.195873 4733 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196000 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196014 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196023 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196033 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196041 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196049 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196057 4733 feature_gate.go:330] unrecognized feature gate: Example Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196068 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196077 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196086 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196094 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196103 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196111 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196118 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196127 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196134 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196142 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196150 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196160 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196170 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196178 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196186 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196194 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196202 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196230 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196238 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196246 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196254 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196262 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196270 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196277 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196287 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196298 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196309 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196347 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196357 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196366 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196375 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196383 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196391 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196399 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196408 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196416 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196425 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196433 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196441 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196451 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196461 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196470 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196480 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196490 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196498 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196506 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196514 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196523 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196532 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196539 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196548 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196556 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196565 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196573 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196582 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196590 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196598 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196607 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196615 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196623 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196631 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196638 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196647 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196654 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.196668 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196908 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196922 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196931 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196939 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196948 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196956 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196964 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196971 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196980 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.196990 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197000 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197009 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197017 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197025 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197034 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197042 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197050 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197057 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197065 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197074 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197082 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197091 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197100 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197108 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197117 4733 feature_gate.go:330] unrecognized feature gate: Example Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197126 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197135 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197143 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197151 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197158 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197167 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197175 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197183 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197191 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197199 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197208 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197215 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197227 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197235 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197243 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197251 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197259 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197268 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197276 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197284 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197292 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197300 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197307 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197315 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197323 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197330 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197338 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197346 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197354 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197362 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197373 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197383 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197392 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197401 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197408 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197416 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197427 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197437 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197448 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197456 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197465 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197475 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197484 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197491 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197499 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.197507 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.197519 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.198837 4733 server.go:940] "Client rotation is on, will bootstrap in background" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.203329 4733 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.203475 4733 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.204428 4733 server.go:997] "Starting client certificate rotation" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.204457 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.204976 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-05 21:22:03.361384543 +0000 UTC Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.205151 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 771h43m5.156241405s for next certificate rotation Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.212214 4733 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.217340 4733 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.228270 4733 log.go:25] "Validated CRI v1 runtime API" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.241597 4733 log.go:25] "Validated CRI v1 image API" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.243325 4733 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.245782 4733 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-04-17-34-16-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.245836 4733 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.263392 4733 manager.go:217] Machine: {Timestamp:2025-12-04 17:38:58.262019997 +0000 UTC m=+0.217381073 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:61d6cd1d-2e94-4088-ae6a-14e04c4bcca0 BootID:d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:5a:2f:8e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:5a:2f:8e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ac:4d:a0 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:43:69:9a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:96:81:68 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d4:ea:fa Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:a6:3c:2f Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4e:d2:a6:41:80:c8 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1e:4a:13:e5:18:9c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.263684 4733 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.263904 4733 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.264186 4733 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.264378 4733 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.264411 4733 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.264627 4733 topology_manager.go:138] "Creating topology manager with none policy" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.264640 4733 container_manager_linux.go:303] "Creating device plugin manager" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.264915 4733 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.264946 4733 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.265264 4733 state_mem.go:36] "Initialized new in-memory state store" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.265465 4733 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.266520 4733 kubelet.go:418] "Attempting to sync node with API server" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.266542 4733 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.266566 4733 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.266580 4733 kubelet.go:324] "Adding apiserver pod source" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.266598 4733 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.268330 4733 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.268691 4733 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.269450 4733 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.269633 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.269639 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.269754 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.269771 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.269988 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270007 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270013 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270020 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270031 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270038 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270044 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270055 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270062 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270069 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270096 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270103 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270293 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.270648 4733 server.go:1280] "Started kubelet" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.271117 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.271193 4733 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.271235 4733 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.272062 4733 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 04 17:38:58 crc systemd[1]: Started Kubernetes Kubelet. Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.273546 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.273568 4733 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.273836 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 00:26:57.067002641 +0000 UTC Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.273485 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.246:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e13dbb162dbaa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 17:38:58.270624682 +0000 UTC m=+0.225985728,LastTimestamp:2025-12-04 17:38:58.270624682 +0000 UTC m=+0.225985728,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.274130 4733 server.go:460] "Adding debug handlers to kubelet server" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.274447 4733 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.274471 4733 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.274629 4733 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.276105 4733 factory.go:55] Registering systemd factory Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.276138 4733 factory.go:221] Registration of the systemd container factory successfully Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.279176 4733 factory.go:153] Registering CRI-O factory Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.279205 4733 factory.go:221] Registration of the crio container factory successfully Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.279268 4733 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.279291 4733 factory.go:103] Registering Raw factory Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.279314 4733 manager.go:1196] Started watching for new ooms in manager Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.280379 4733 manager.go:319] Starting recovery of all containers Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.282556 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.282916 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.285992 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.286608 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="200ms" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.295243 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.295351 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.295383 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.295406 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.295428 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.295494 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.295515 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296040 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296096 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296121 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296215 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296241 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296267 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296298 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296324 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296348 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296377 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296405 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296432 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296458 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296484 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296510 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296535 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296560 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296605 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296633 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296668 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296745 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296774 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296863 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296894 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296921 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296949 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.296978 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297003 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297027 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297052 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297075 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297101 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297131 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297196 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297222 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297246 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297270 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297293 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297311 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297329 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297346 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297367 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297385 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297406 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297425 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297452 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297473 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297493 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297513 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297532 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297550 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297568 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297585 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297604 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297623 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297642 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297660 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297687 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297706 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297724 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297743 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297762 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297782 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297831 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297850 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297869 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297886 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297905 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297924 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.297942 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.300413 4733 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.300539 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.300673 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.300776 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.300881 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.300958 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301043 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301119 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301190 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301279 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301376 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301404 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301420 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301856 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301887 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301901 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301915 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301931 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301949 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301963 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301979 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.301994 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302009 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302022 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302035 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302048 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302104 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302120 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302140 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302158 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302174 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302190 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302205 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302219 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302234 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302247 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302260 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302273 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302288 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302301 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302313 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302326 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302338 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302353 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302367 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302380 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302392 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302404 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302417 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302431 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302444 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302457 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302470 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302482 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302499 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302513 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302526 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302538 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302551 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302587 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302600 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302614 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302627 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302640 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302653 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302667 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302680 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302694 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302707 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302720 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302760 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302772 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302822 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302836 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302849 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302864 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302877 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302919 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302932 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302946 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302959 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302972 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302985 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.302998 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303012 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303024 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303038 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303050 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303063 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303075 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303087 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303101 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303116 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303129 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303141 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303154 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303168 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303180 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303194 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303207 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303222 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303234 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303247 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303259 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303272 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303284 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303296 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303309 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303322 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303334 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303347 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303363 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303376 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303389 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303404 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303415 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303427 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303440 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303469 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303481 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303493 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303504 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303516 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303528 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303540 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303552 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303563 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303574 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303588 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303600 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303612 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303625 4733 reconstruct.go:97] "Volume reconstruction finished" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.303634 4733 reconciler.go:26] "Reconciler: start to sync state" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.305418 4733 manager.go:324] Recovery completed Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.320488 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.328602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.328842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.328948 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.329908 4733 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.329999 4733 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.330082 4733 state_mem.go:36] "Initialized new in-memory state store" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.332113 4733 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.334046 4733 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.334101 4733 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.334165 4733 kubelet.go:2335] "Starting kubelet main sync loop" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.334242 4733 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.335858 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.335992 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.341055 4733 policy_none.go:49] "None policy: Start" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.342238 4733 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.342372 4733 state_mem.go:35] "Initializing new in-memory state store" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.386657 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.410779 4733 manager.go:334] "Starting Device Plugin manager" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.411044 4733 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.411064 4733 server.go:79] "Starting device plugin registration server" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.411587 4733 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.411613 4733 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.412007 4733 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.412086 4733 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.412096 4733 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.423056 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.435265 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.435319 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.436292 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.436347 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.436361 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.436540 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.436919 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.436968 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437673 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437835 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437877 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437943 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.437961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.438239 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.438267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.438278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.438383 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.438592 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.438646 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439260 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439362 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439416 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439485 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439515 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439652 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.439669 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.440295 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.440314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.440317 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.440336 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.440346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.440321 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.440597 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.440630 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.441730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.441756 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.441767 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.458179 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.246:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e13dbb162dbaa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 17:38:58.270624682 +0000 UTC m=+0.225985728,LastTimestamp:2025-12-04 17:38:58.270624682 +0000 UTC m=+0.225985728,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.487959 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="400ms" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.506534 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.506642 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.506708 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.506768 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.506838 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.506911 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.506965 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.507024 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.507074 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.507174 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.507211 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.507252 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.507320 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.507398 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.507428 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.511966 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.513350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.513394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.513445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.513472 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.514358 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608268 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608345 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608424 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608511 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608545 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608577 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608667 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608614 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608686 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608729 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608765 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608594 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608850 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608870 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608886 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608917 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608924 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608925 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608958 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.608973 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.609004 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.609007 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.609030 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.609053 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.609083 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.609121 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.609275 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.715285 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.716744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.716818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.716835 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.716868 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.717334 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.788694 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.813663 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.819572 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-c78169222f2e93d5021772d2bea0436c5048e08311a597631259be205045994a WatchSource:0}: Error finding container c78169222f2e93d5021772d2bea0436c5048e08311a597631259be205045994a: Status 404 returned error can't find the container with id c78169222f2e93d5021772d2bea0436c5048e08311a597631259be205045994a Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.823182 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.846117 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.849293 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-e54983923be1ab1f58b766bea6f67347219a4b34c0fdd93d97855717ca57f446 WatchSource:0}: Error finding container e54983923be1ab1f58b766bea6f67347219a4b34c0fdd93d97855717ca57f446: Status 404 returned error can't find the container with id e54983923be1ab1f58b766bea6f67347219a4b34c0fdd93d97855717ca57f446 Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.854206 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-0f7525f860d24394fecee6db2698cd3a84581eb6dc0da4f06d31c7aae50208e9 WatchSource:0}: Error finding container 0f7525f860d24394fecee6db2698cd3a84581eb6dc0da4f06d31c7aae50208e9: Status 404 returned error can't find the container with id 0f7525f860d24394fecee6db2698cd3a84581eb6dc0da4f06d31c7aae50208e9 Dec 04 17:38:58 crc kubenswrapper[4733]: I1204 17:38:58.856846 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.882470 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-0af4f0d0982d4f8f318592468ca7774472de8d2b0837958fd8b2b7c45f53fa69 WatchSource:0}: Error finding container 0af4f0d0982d4f8f318592468ca7774472de8d2b0837958fd8b2b7c45f53fa69: Status 404 returned error can't find the container with id 0af4f0d0982d4f8f318592468ca7774472de8d2b0837958fd8b2b7c45f53fa69 Dec 04 17:38:58 crc kubenswrapper[4733]: W1204 17:38:58.884336 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-6af437d5d17e058fbc49d790d552e4bd4486f070f6ce16b4f41f1467716b9a19 WatchSource:0}: Error finding container 6af437d5d17e058fbc49d790d552e4bd4486f070f6ce16b4f41f1467716b9a19: Status 404 returned error can't find the container with id 6af437d5d17e058fbc49d790d552e4bd4486f070f6ce16b4f41f1467716b9a19 Dec 04 17:38:58 crc kubenswrapper[4733]: E1204 17:38:58.889171 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="800ms" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.118186 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.119119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.119154 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.119166 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.119188 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 17:38:59 crc kubenswrapper[4733]: E1204 17:38:59.119596 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Dec 04 17:38:59 crc kubenswrapper[4733]: W1204 17:38:59.213118 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:59 crc kubenswrapper[4733]: E1204 17:38:59.213193 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.272393 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.274456 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 04:34:39.743484361 +0000 UTC Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.341787 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7" exitCode=0 Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.341879 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.342066 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0f7525f860d24394fecee6db2698cd3a84581eb6dc0da4f06d31c7aae50208e9"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.342216 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.343699 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590" exitCode=0 Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.343733 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.343775 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e54983923be1ab1f58b766bea6f67347219a4b34c0fdd93d97855717ca57f446"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.343964 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.344217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.344250 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.344259 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.345281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.345319 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.345328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.349381 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.350141 4733 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="009385cae8fd134188bcaa730654e0f8685230315a6b5efae2b6f9b707678a05" exitCode=0 Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.350201 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"009385cae8fd134188bcaa730654e0f8685230315a6b5efae2b6f9b707678a05"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.350265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c78169222f2e93d5021772d2bea0436c5048e08311a597631259be205045994a"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.350367 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.351635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.351671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.351685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.352531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.352578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.352598 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.354185 4733 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660" exitCode=0 Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.354239 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.354281 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6af437d5d17e058fbc49d790d552e4bd4486f070f6ce16b4f41f1467716b9a19"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.354437 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.355568 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.355596 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0af4f0d0982d4f8f318592468ca7774472de8d2b0837958fd8b2b7c45f53fa69"} Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.355719 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.355740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.355750 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:59 crc kubenswrapper[4733]: W1204 17:38:59.372781 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:59 crc kubenswrapper[4733]: E1204 17:38:59.372923 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Dec 04 17:38:59 crc kubenswrapper[4733]: W1204 17:38:59.415458 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:59 crc kubenswrapper[4733]: E1204 17:38:59.415571 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Dec 04 17:38:59 crc kubenswrapper[4733]: W1204 17:38:59.677065 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Dec 04 17:38:59 crc kubenswrapper[4733]: E1204 17:38:59.677130 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Dec 04 17:38:59 crc kubenswrapper[4733]: E1204 17:38:59.689838 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="1.6s" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.920138 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.921449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.921486 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.921501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:38:59 crc kubenswrapper[4733]: I1204 17:38:59.921526 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 17:38:59 crc kubenswrapper[4733]: E1204 17:38:59.921955 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.274866 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 11:43:25.396635211 +0000 UTC Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.274928 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1050h4m25.121709607s for next certificate rotation Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.359826 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.359871 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.359887 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.359980 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.360869 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.360894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.360905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.369238 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.369276 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.369283 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.369380 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.370551 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.370584 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.370597 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.374689 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.374718 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.374732 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.376716 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c" exitCode=0 Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.376754 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.376858 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.377737 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.377760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.377768 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.379537 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5436fb6b7414b11a3dcfae0bfa40d49a0fa9fb9924277710d10e6797412f8543"} Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.379614 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.380307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.380342 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:00 crc kubenswrapper[4733]: I1204 17:39:00.380352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.387762 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581"} Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.387894 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.387893 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718"} Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.389124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.389175 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.389193 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.391718 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d" exitCode=0 Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.391880 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.392566 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d"} Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.392705 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.392998 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.393346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.393526 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.393701 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.393734 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.394025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.394047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.394107 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.394156 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.394175 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.522677 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.524862 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.524931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.524954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:01 crc kubenswrapper[4733]: I1204 17:39:01.525003 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.113411 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.372828 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.398324 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243"} Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.398394 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407"} Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.398418 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca"} Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.398353 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.398477 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.399788 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.399838 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.399852 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.399854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.399901 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.399920 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:02 crc kubenswrapper[4733]: I1204 17:39:02.459066 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.030736 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.226852 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.235423 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.282542 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.282716 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.284558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.284619 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.284637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.407597 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550"} Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.407662 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.407682 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d"} Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.407730 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.407779 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409569 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409618 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409575 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:03 crc kubenswrapper[4733]: I1204 17:39:03.409848 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.410054 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.410097 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.410152 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.410080 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.411773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.411810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.411821 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.411926 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.411971 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.411998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.412422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.412449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:04 crc kubenswrapper[4733]: I1204 17:39:04.412458 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:05 crc kubenswrapper[4733]: I1204 17:39:05.372986 4733 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 17:39:05 crc kubenswrapper[4733]: I1204 17:39:05.373087 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 17:39:05 crc kubenswrapper[4733]: I1204 17:39:05.410201 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 04 17:39:05 crc kubenswrapper[4733]: I1204 17:39:05.412096 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:05 crc kubenswrapper[4733]: I1204 17:39:05.413428 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:05 crc kubenswrapper[4733]: I1204 17:39:05.413468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:05 crc kubenswrapper[4733]: I1204 17:39:05.413485 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.032855 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.032996 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.034046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.034109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.034132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.078220 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.419972 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.421847 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.421983 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:08 crc kubenswrapper[4733]: I1204 17:39:08.422014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:08 crc kubenswrapper[4733]: E1204 17:39:08.423163 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 17:39:10 crc kubenswrapper[4733]: I1204 17:39:10.280341 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 04 17:39:10 crc kubenswrapper[4733]: I1204 17:39:10.730987 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 04 17:39:10 crc kubenswrapper[4733]: I1204 17:39:10.731225 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:10 crc kubenswrapper[4733]: I1204 17:39:10.732665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:10 crc kubenswrapper[4733]: I1204 17:39:10.732741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:10 crc kubenswrapper[4733]: I1204 17:39:10.732759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:10 crc kubenswrapper[4733]: W1204 17:39:10.906862 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 17:39:10 crc kubenswrapper[4733]: I1204 17:39:10.906984 4733 trace.go:236] Trace[1342648620]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 17:39:00.905) (total time: 10001ms): Dec 04 17:39:10 crc kubenswrapper[4733]: Trace[1342648620]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (17:39:10.906) Dec 04 17:39:10 crc kubenswrapper[4733]: Trace[1342648620]: [10.001842914s] [10.001842914s] END Dec 04 17:39:10 crc kubenswrapper[4733]: E1204 17:39:10.907012 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 17:39:11 crc kubenswrapper[4733]: E1204 17:39:11.291555 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 04 17:39:11 crc kubenswrapper[4733]: I1204 17:39:11.329960 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 17:39:11 crc kubenswrapper[4733]: I1204 17:39:11.330032 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 17:39:11 crc kubenswrapper[4733]: I1204 17:39:11.337358 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 17:39:11 crc kubenswrapper[4733]: I1204 17:39:11.337422 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 17:39:12 crc kubenswrapper[4733]: I1204 17:39:12.466853 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]log ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]etcd ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/generic-apiserver-start-informers ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/priority-and-fairness-filter ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-apiextensions-informers ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-apiextensions-controllers ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/crd-informer-synced ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-system-namespaces-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 04 17:39:12 crc kubenswrapper[4733]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/bootstrap-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/start-kube-aggregator-informers ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/apiservice-registration-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/apiservice-discovery-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]autoregister-completion ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/apiservice-openapi-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 04 17:39:12 crc kubenswrapper[4733]: livez check failed Dec 04 17:39:12 crc kubenswrapper[4733]: I1204 17:39:12.466947 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:39:15 crc kubenswrapper[4733]: I1204 17:39:15.374199 4733 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 17:39:15 crc kubenswrapper[4733]: I1204 17:39:15.374302 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 17:39:15 crc kubenswrapper[4733]: I1204 17:39:15.496583 4733 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.317457 4733 trace.go:236] Trace[752499858]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 17:39:02.060) (total time: 14256ms): Dec 04 17:39:16 crc kubenswrapper[4733]: Trace[752499858]: ---"Objects listed" error: 14256ms (17:39:16.317) Dec 04 17:39:16 crc kubenswrapper[4733]: Trace[752499858]: [14.256586609s] [14.256586609s] END Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.317498 4733 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.321947 4733 trace.go:236] Trace[643964383]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 17:39:01.728) (total time: 14593ms): Dec 04 17:39:16 crc kubenswrapper[4733]: Trace[643964383]: ---"Objects listed" error: 14593ms (17:39:16.321) Dec 04 17:39:16 crc kubenswrapper[4733]: Trace[643964383]: [14.593574756s] [14.593574756s] END Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.321983 4733 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.324099 4733 trace.go:236] Trace[1844392116]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 17:39:02.257) (total time: 14066ms): Dec 04 17:39:16 crc kubenswrapper[4733]: Trace[1844392116]: ---"Objects listed" error: 14066ms (17:39:16.324) Dec 04 17:39:16 crc kubenswrapper[4733]: Trace[1844392116]: [14.066184524s] [14.066184524s] END Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.324125 4733 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.325949 4733 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 04 17:39:16 crc kubenswrapper[4733]: E1204 17:39:16.326744 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.365701 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:40026->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.367179 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:40026->192.168.126.11:17697: read: connection reset by peer" Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.442621 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.444676 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718" exitCode=255 Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.444729 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718"} Dec 04 17:39:16 crc kubenswrapper[4733]: I1204 17:39:16.479162 4733 scope.go:117] "RemoveContainer" containerID="e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.277917 4733 apiserver.go:52] "Watching apiserver" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.281538 4733 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.281870 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.282430 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.282494 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.282570 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.282898 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.282914 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.283104 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.283132 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.283385 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.283413 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.285568 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.286170 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.286591 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.286764 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.286955 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.286780 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.289768 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.290973 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.291199 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.327129 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.345885 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.361851 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.377506 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.382278 4733 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.399378 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.414980 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432287 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432373 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432450 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432484 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432517 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432560 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432591 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432654 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432689 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432720 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432749 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432780 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432837 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432870 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432903 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432933 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432967 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.432998 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433028 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433059 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433091 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433133 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433175 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433161 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433208 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433314 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433365 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433418 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433465 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433499 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433530 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433564 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433594 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433625 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433660 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433651 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433698 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433732 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433723 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433899 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433939 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.433971 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434004 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434037 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434069 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434112 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434160 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434205 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434237 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434268 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434300 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434333 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434366 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434397 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434443 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434393 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434475 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434581 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434592 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434623 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434640 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434660 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434669 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434742 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434775 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434837 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434871 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434901 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434935 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434939 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.434967 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435000 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435021 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435034 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435170 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435208 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435247 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435280 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435311 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435339 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435347 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435503 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435511 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435614 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435628 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435651 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435687 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435903 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435945 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435957 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.435977 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436009 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436044 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436086 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436120 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436153 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436161 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436186 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436219 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436250 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436281 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436312 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436349 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436398 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436430 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436464 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436496 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436529 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436560 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436593 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436629 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436666 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436715 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436763 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436843 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436892 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436938 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436987 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437034 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437097 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437147 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437193 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437242 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437287 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437322 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437354 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437387 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437425 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437459 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437491 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437526 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437562 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437611 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437657 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437696 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437733 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437769 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437847 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437893 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437925 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437963 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437995 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438029 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438062 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438096 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438128 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438162 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438196 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438231 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438265 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438300 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438339 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438374 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438466 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438500 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438535 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438569 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438602 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438634 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438671 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438706 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438741 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438775 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438870 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438910 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438944 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438980 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439026 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439061 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439211 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439251 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439286 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439322 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439442 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439504 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439665 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439719 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439768 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439870 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439923 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439962 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440001 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440036 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440074 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440109 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440145 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440179 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440214 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440247 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440300 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440341 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440376 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440409 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440443 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440480 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440516 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440551 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440587 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440620 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440653 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440693 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440743 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440778 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440858 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440900 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440939 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440989 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441040 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441135 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441196 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441255 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441314 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441368 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441420 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441525 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441570 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441627 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441678 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441733 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441789 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441877 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441989 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442023 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442052 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442080 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442103 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442127 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442149 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442172 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442193 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442213 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442233 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442253 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442274 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442295 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.450186 4733 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.456883 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.457566 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.461455 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.462144 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.462432 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.464539 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.436588 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437121 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437223 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437660 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437679 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437765 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.437987 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438189 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438214 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438602 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438720 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438741 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.438955 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439272 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439386 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439845 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.475127 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439921 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.439954 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440008 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440270 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440449 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440657 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440687 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.440946 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.441382 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442007 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442179 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.442396 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.443075 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.443075 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.443695 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.443908 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:39:17.943877395 +0000 UTC m=+19.899238481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.444054 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.444259 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.444484 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.444516 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.444578 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.444589 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.444771 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.444929 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.445044 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.445271 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.445523 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.445615 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.445680 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.445970 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.446091 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.446651 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.447031 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.447912 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.448066 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.448120 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.448227 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.448248 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.448009 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.448652 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449007 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449044 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449111 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449103 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449170 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449451 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449506 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449702 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.449997 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.450202 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.450254 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.450336 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.450380 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.451439 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.451465 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.451624 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.451753 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.451369 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.452184 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.452511 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.452746 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.453266 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.453335 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.453372 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.453399 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.453415 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.453636 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.453900 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.453973 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.454214 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.454287 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.454363 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.454418 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.454462 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.454754 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.455557 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.455709 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.456457 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.456051 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.457049 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.457514 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.458334 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.459230 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.460084 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.460205 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.460712 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.460992 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.462241 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.464202 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.463537 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.465658 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.466551 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.466597 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.466948 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.467614 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.467754 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.468300 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.468330 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.468704 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.468821 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.469147 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.469452 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.469746 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.470082 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.470962 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.471082 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.471595 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.471614 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.472621 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.474921 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.476140 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.477594 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.478466 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.478948 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.479315 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.481783 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.481975 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.482338 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.482783 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.483401 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.483531 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.483596 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.483849 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.484113 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:17.984067027 +0000 UTC m=+19.939428163 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.484190 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:17.984163236 +0000 UTC m=+19.939524312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.484265 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.485003 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.485434 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.485598 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.485907 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.486050 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.483535 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.486352 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.486840 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.487046 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.487272 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.487518 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.487863 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.488299 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.488661 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.489003 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.490061 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.490407 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c"} Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.491006 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.495064 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.495138 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.495158 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.495292 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:17.995264803 +0000 UTC m=+19.950625939 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.487345 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.495957 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.496216 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.497517 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.498210 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.501575 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.502537 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.503140 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.504369 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.504390 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.504406 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.504457 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:18.004441434 +0000 UTC m=+19.959802490 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.504821 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.504988 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.505340 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.505834 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.506072 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.506456 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.506573 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.508407 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.508541 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.511340 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.511569 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.511558 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.512936 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.513090 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.513446 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.514302 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.514397 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.514347 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.514445 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.514581 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.514620 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.520366 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.524026 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.531001 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.535861 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.539303 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.543688 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.543761 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.543911 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.543927 4733 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.543940 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.543952 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.543989 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544010 4733 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544020 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544031 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544043 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544081 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544113 4733 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544125 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544161 4733 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.543936 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544172 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544264 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544276 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544291 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544303 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544314 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544326 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544337 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544348 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544361 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544372 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544383 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544394 4733 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544405 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544415 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544425 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544436 4733 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544448 4733 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544460 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544470 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544481 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544492 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544503 4733 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544513 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544524 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544535 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544545 4733 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544556 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544567 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544578 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544588 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544599 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544609 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544620 4733 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544630 4733 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544640 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544651 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544660 4733 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544671 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544682 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544693 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544703 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544717 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544728 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544739 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544750 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544760 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544772 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544782 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544810 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544823 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544834 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544846 4733 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544860 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544871 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544883 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544895 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544906 4733 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544917 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544928 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544939 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544949 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544961 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544972 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544983 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.544993 4733 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545004 4733 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545015 4733 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545025 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545036 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545048 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545059 4733 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545072 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545083 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545094 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545106 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545117 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545128 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545139 4733 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545193 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545204 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545215 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545226 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545238 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545249 4733 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545259 4733 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545269 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545279 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545290 4733 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545302 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545313 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545324 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545335 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545347 4733 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545359 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545369 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545380 4733 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545391 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545402 4733 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545414 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545425 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545436 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545447 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545458 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545469 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545480 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545490 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545500 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545510 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545521 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545531 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545542 4733 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545552 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545563 4733 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545576 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545586 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545597 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545608 4733 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545619 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545630 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545642 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545651 4733 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545661 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545672 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545682 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545693 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545703 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545713 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545723 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545734 4733 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545744 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545757 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545767 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545778 4733 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545789 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545817 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545828 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545839 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545850 4733 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545861 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545871 4733 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545882 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545892 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545903 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545913 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545924 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545935 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545946 4733 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545956 4733 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545966 4733 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545977 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.545990 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546001 4733 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546012 4733 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546024 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546035 4733 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546045 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546056 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546066 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546077 4733 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546088 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546100 4733 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546115 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546126 4733 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546137 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546153 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.546178 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.548700 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.554341 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.556921 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.572080 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.583661 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.596643 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.597334 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.605158 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.617598 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.647170 4733 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 17:39:17 crc kubenswrapper[4733]: I1204 17:39:17.949300 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:17 crc kubenswrapper[4733]: E1204 17:39:17.949529 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:39:18.949491529 +0000 UTC m=+20.904852595 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.041134 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.050636 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.050685 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.050738 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.050765 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.050888 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.050937 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:19.050922231 +0000 UTC m=+21.006283287 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051181 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051218 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:19.051209267 +0000 UTC m=+21.006570313 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051279 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051293 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051305 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051333 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:19.051324026 +0000 UTC m=+21.006685072 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051391 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051405 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051414 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.051440 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:19.051431874 +0000 UTC m=+21.006792930 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.051771 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.055941 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.069700 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.088831 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.112769 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.132874 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.145269 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.160910 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.341482 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.343026 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.344458 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.346529 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.347660 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.349530 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.350875 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.351468 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.352862 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.354125 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.356033 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.357123 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.359244 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.360310 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.361430 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.363264 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.364307 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.366253 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.367048 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.368320 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.369084 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.370359 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.371273 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.372506 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.374277 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.375112 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.376138 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.376925 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.378337 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.378952 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.380143 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.380749 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.381415 4733 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.382017 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.384101 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.384769 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.385826 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.387748 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.388583 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.389475 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.389886 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.390853 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.392445 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.393258 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.394639 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.395842 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.397170 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.397750 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.398935 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.399574 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.400970 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.401558 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.402617 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.403220 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.403914 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.405173 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.405759 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.406381 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.431998 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.460135 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.482634 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.495886 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e"} Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.495953 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326"} Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.495979 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6dbfe888e4c6ca410785cd9d5d8d2be2bfe038149467e25fdeace6348331aa6d"} Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.498221 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6a9078afd9641644906a2b980230e956ff544ece7ac2fc27254ec0df00e5f2b5"} Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.501870 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334"} Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.501919 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e364ddbd5793a85ebf8469430cca320d813fba06a4361aa54458a270f78fe7d5"} Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.503562 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.527299 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.545858 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.563263 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.582114 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.602573 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.631124 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.650842 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.673770 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:18 crc kubenswrapper[4733]: I1204 17:39:18.960762 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:18 crc kubenswrapper[4733]: E1204 17:39:18.960958 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:39:20.960939792 +0000 UTC m=+22.916300848 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.061748 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.061942 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.061956 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.061977 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.061996 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.062007 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.062054 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062073 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:21.062052796 +0000 UTC m=+23.017413872 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062157 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062170 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062243 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:21.062223011 +0000 UTC m=+23.017584067 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062268 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:21.062258642 +0000 UTC m=+23.017619698 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062350 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062365 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062378 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.062405 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:21.062397586 +0000 UTC m=+23.017758642 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.334603 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.334786 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.335354 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.335479 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.335565 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.335665 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.527082 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.528641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.528716 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.528736 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.528861 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.537717 4733 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.538175 4733 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.539708 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.539763 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.539787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.539841 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.539861 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.575331 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.580338 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.580396 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.580414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.580437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.580455 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.600363 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.604642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.604715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.604735 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.605174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.605236 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.626039 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.632076 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.632139 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.632157 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.632181 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.632197 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.654407 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.662655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.662714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.662731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.662756 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.662772 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.683988 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:19 crc kubenswrapper[4733]: E1204 17:39:19.684226 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.686471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.686523 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.686540 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.686567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.686584 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.789592 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.789884 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.790048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.790188 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.790306 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.894018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.894123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.894152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.894184 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.894207 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.997180 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.997242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.997260 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.997287 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:19 crc kubenswrapper[4733]: I1204 17:39:19.997304 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:19Z","lastTransitionTime":"2025-12-04T17:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.100426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.100477 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.100495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.100518 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.100534 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.203061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.203089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.203099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.203114 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.203126 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.305868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.305944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.305969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.306000 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.306021 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.409698 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.409761 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.409784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.409852 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.409876 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.512730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.512849 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.512875 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.512906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.512927 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.615876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.615938 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.615956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.615981 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.616001 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.719520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.719577 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.719595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.719618 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.719638 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.769538 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.789139 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.791439 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.793394 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.815123 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.822830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.822880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.822901 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.822929 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.822978 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.836259 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.851500 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.870886 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.887108 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.903875 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.926229 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.926464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.926545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.926654 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.926738 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:20Z","lastTransitionTime":"2025-12-04T17:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.927922 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.954365 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.969377 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.982043 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:20 crc kubenswrapper[4733]: E1204 17:39:20.982244 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:39:24.982212138 +0000 UTC m=+26.937573214 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:39:20 crc kubenswrapper[4733]: I1204 17:39:20.984406 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:20Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.021552 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.030399 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.030446 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.030458 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.030474 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.030485 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.043344 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.062308 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.075948 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.083053 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.083124 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.083163 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.083201 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083341 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083364 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083375 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083404 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083439 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083474 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:25.083444456 +0000 UTC m=+27.038805542 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083480 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083511 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083513 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:25.083495118 +0000 UTC m=+27.038856204 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083512 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083689 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:25.083647042 +0000 UTC m=+27.039008128 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.083746 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:25.083720734 +0000 UTC m=+27.039081820 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.093392 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.112174 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.133123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.133280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.133367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.133455 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.133533 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.236825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.236869 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.236883 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.236903 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.236915 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.334692 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.334725 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.334749 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.334943 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.335083 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:21 crc kubenswrapper[4733]: E1204 17:39:21.335260 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.341657 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.341721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.341741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.341772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.341823 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.444548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.445034 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.445112 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.445203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.445299 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.511870 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.531787 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.548239 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.548290 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.548307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.548329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.548345 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.564488 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.586700 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.607448 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.634231 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.651577 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.651637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.651658 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.651684 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.651702 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.653476 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.673260 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.692735 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.713849 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.755158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.755229 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.755254 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.755283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.755307 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.858550 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.858624 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.858648 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.858676 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.858697 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.961602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.961982 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.962167 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.962337 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:21 crc kubenswrapper[4733]: I1204 17:39:21.962494 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:21Z","lastTransitionTime":"2025-12-04T17:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.065691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.065754 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.065771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.065831 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.065851 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.169146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.169190 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.169202 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.169219 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.169232 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.275377 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.275441 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.275459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.275485 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.275510 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.377932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.378253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.378427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.378602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.378747 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.380081 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.385910 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.401548 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.424825 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.445020 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.465483 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.481599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.481655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.481673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.481697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.481716 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.485162 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.519647 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.541388 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.563638 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.584358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.584471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.584491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.584515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.584535 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.586141 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.604741 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.624843 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.651872 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.671169 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.687587 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.687902 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.688061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.688209 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.688370 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.694578 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.736309 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.759963 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.780754 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.791194 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.791433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.791576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.791745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.791920 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.801397 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:22Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.895246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.895312 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.895329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.895354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.895374 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.998498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.998544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.998560 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.998583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:22 crc kubenswrapper[4733]: I1204 17:39:22.998600 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:22Z","lastTransitionTime":"2025-12-04T17:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.101141 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.101190 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.101201 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.101218 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.101234 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.204459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.204753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.204904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.205038 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.205187 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.308269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.308304 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.308340 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.308355 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.308365 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.335149 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.335228 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:23 crc kubenswrapper[4733]: E1204 17:39:23.335289 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:23 crc kubenswrapper[4733]: E1204 17:39:23.335361 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.335178 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:23 crc kubenswrapper[4733]: E1204 17:39:23.335650 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.410583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.410843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.410928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.411007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.411093 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.512907 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.512940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.512948 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.512961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.512970 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.614945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.615022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.615032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.615047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.615057 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.717536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.717584 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.717600 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.717618 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.717633 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.820248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.820310 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.820328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.820352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.820370 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.923252 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.923291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.923301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.923315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:23 crc kubenswrapper[4733]: I1204 17:39:23.923324 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:23Z","lastTransitionTime":"2025-12-04T17:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.025821 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.025868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.025879 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.025900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.025913 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.128581 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.128616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.128627 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.128646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.128657 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.199088 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-xf8lf"] Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.199748 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-w7n9m"] Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.199916 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.200057 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.200061 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-472m8"] Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.200573 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.203636 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.203772 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-6tl82"] Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.204057 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6tl82" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.204130 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.204278 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.204536 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.206103 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.206215 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.208377 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.208389 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.208560 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.208623 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.208726 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.209531 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.211174 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.211209 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.211523 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.230205 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.230282 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.230305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.230312 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.230325 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.230336 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.259909 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.276390 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.288581 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.304359 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.316694 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-cnibin\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.316779 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.316849 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-socket-dir-parent\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.316872 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8ef97472-7ccc-4d3d-a127-e70bced45051-cni-binary-copy\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.316920 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pcsn\" (UniqueName: \"kubernetes.io/projected/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-kube-api-access-9pcsn\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.316936 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-conf-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.316952 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-cnibin\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317044 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-cni-multus\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317062 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8ef97472-7ccc-4d3d-a127-e70bced45051-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317128 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-k8s-cni-cncf-io\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317146 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-etc-kubernetes\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317223 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/829fda7a-6a96-4007-b2f5-0be03782a764-cni-binary-copy\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317242 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-kubelet\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317302 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwgwr\" (UniqueName: \"kubernetes.io/projected/21f90af6-f871-4cce-88b4-824c4a15e42c-kube-api-access-gwgwr\") pod \"node-resolver-6tl82\" (UID: \"21f90af6-f871-4cce-88b4-824c4a15e42c\") " pod="openshift-dns/node-resolver-6tl82" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317327 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-rootfs\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317401 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-mcd-auth-proxy-config\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317422 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-multus-certs\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317480 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-os-release\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317543 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-cni-bin\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317558 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-hostroot\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317578 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-system-cni-dir\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317624 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-proxy-tls\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317641 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-cni-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317696 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj56j\" (UniqueName: \"kubernetes.io/projected/8ef97472-7ccc-4d3d-a127-e70bced45051-kube-api-access-bj56j\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317718 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-netns\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317733 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/21f90af6-f871-4cce-88b4-824c4a15e42c-hosts-file\") pod \"node-resolver-6tl82\" (UID: \"21f90af6-f871-4cce-88b4-824c4a15e42c\") " pod="openshift-dns/node-resolver-6tl82" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317778 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swzhg\" (UniqueName: \"kubernetes.io/projected/829fda7a-6a96-4007-b2f5-0be03782a764-kube-api-access-swzhg\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317829 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-system-cni-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317846 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/829fda7a-6a96-4007-b2f5-0be03782a764-multus-daemon-config\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.317911 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-os-release\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.327387 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.332884 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.333043 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.333101 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.333165 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.333218 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.339728 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.350370 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.362473 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.372247 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.381527 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.393620 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.406456 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418806 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/829fda7a-6a96-4007-b2f5-0be03782a764-cni-binary-copy\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418854 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-kubelet\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418873 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwgwr\" (UniqueName: \"kubernetes.io/projected/21f90af6-f871-4cce-88b4-824c4a15e42c-kube-api-access-gwgwr\") pod \"node-resolver-6tl82\" (UID: \"21f90af6-f871-4cce-88b4-824c4a15e42c\") " pod="openshift-dns/node-resolver-6tl82" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418901 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-rootfs\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418919 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-mcd-auth-proxy-config\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418935 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-multus-certs\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418950 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-os-release\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418971 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-cni-bin\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418984 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-hostroot\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.418998 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-system-cni-dir\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419012 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-proxy-tls\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419025 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-cni-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419052 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj56j\" (UniqueName: \"kubernetes.io/projected/8ef97472-7ccc-4d3d-a127-e70bced45051-kube-api-access-bj56j\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419075 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-netns\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419089 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/21f90af6-f871-4cce-88b4-824c4a15e42c-hosts-file\") pod \"node-resolver-6tl82\" (UID: \"21f90af6-f871-4cce-88b4-824c4a15e42c\") " pod="openshift-dns/node-resolver-6tl82" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419102 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swzhg\" (UniqueName: \"kubernetes.io/projected/829fda7a-6a96-4007-b2f5-0be03782a764-kube-api-access-swzhg\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419117 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-system-cni-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419096 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-kubelet\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419133 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/829fda7a-6a96-4007-b2f5-0be03782a764-multus-daemon-config\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419207 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-system-cni-dir\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419249 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-rootfs\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419254 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-os-release\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419327 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-cnibin\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419361 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419395 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-socket-dir-parent\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419423 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8ef97472-7ccc-4d3d-a127-e70bced45051-cni-binary-copy\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419454 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pcsn\" (UniqueName: \"kubernetes.io/projected/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-kube-api-access-9pcsn\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419482 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-conf-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419510 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-cnibin\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419539 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-cni-multus\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419570 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8ef97472-7ccc-4d3d-a127-e70bced45051-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419619 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-k8s-cni-cncf-io\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419649 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-etc-kubernetes\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419732 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-etc-kubernetes\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419772 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-mcd-auth-proxy-config\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419774 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/829fda7a-6a96-4007-b2f5-0be03782a764-multus-daemon-config\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419833 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-multus-certs\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419852 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-cnibin\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419861 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-os-release\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419879 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-os-release\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419904 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-cni-bin\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419925 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-hostroot\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.419924 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-netns\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420068 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/21f90af6-f871-4cce-88b4-824c4a15e42c-hosts-file\") pod \"node-resolver-6tl82\" (UID: \"21f90af6-f871-4cce-88b4-824c4a15e42c\") " pod="openshift-dns/node-resolver-6tl82" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420103 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-cnibin\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420103 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-cni-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420158 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-socket-dir-parent\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420235 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-var-lib-cni-multus\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420441 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-system-cni-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420523 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-host-run-k8s-cni-cncf-io\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420572 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8ef97472-7ccc-4d3d-a127-e70bced45051-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420581 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/829fda7a-6a96-4007-b2f5-0be03782a764-multus-conf-dir\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420595 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8ef97472-7ccc-4d3d-a127-e70bced45051-cni-binary-copy\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.420991 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8ef97472-7ccc-4d3d-a127-e70bced45051-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.421179 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/829fda7a-6a96-4007-b2f5-0be03782a764-cni-binary-copy\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.423015 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.426512 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-proxy-tls\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.435709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.436051 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.436064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.436082 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.436099 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.436675 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swzhg\" (UniqueName: \"kubernetes.io/projected/829fda7a-6a96-4007-b2f5-0be03782a764-kube-api-access-swzhg\") pod \"multus-472m8\" (UID: \"829fda7a-6a96-4007-b2f5-0be03782a764\") " pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.438174 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj56j\" (UniqueName: \"kubernetes.io/projected/8ef97472-7ccc-4d3d-a127-e70bced45051-kube-api-access-bj56j\") pod \"multus-additional-cni-plugins-xf8lf\" (UID: \"8ef97472-7ccc-4d3d-a127-e70bced45051\") " pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.444370 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwgwr\" (UniqueName: \"kubernetes.io/projected/21f90af6-f871-4cce-88b4-824c4a15e42c-kube-api-access-gwgwr\") pod \"node-resolver-6tl82\" (UID: \"21f90af6-f871-4cce-88b4-824c4a15e42c\") " pod="openshift-dns/node-resolver-6tl82" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.451766 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pcsn\" (UniqueName: \"kubernetes.io/projected/5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9-kube-api-access-9pcsn\") pod \"machine-config-daemon-w7n9m\" (UID: \"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\") " pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.457282 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.476921 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.494184 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.511846 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.512097 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.519633 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:39:24 crc kubenswrapper[4733]: W1204 17:39:24.524576 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ef97472_7ccc_4d3d_a127_e70bced45051.slice/crio-8558088948fb7eb1c7d802bf5a61d5772dfb58586f2ba88eb87c2d292ee27a8d WatchSource:0}: Error finding container 8558088948fb7eb1c7d802bf5a61d5772dfb58586f2ba88eb87c2d292ee27a8d: Status 404 returned error can't find the container with id 8558088948fb7eb1c7d802bf5a61d5772dfb58586f2ba88eb87c2d292ee27a8d Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.526754 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-472m8" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.530829 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.534080 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6tl82" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.538873 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.538900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.538913 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.538929 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.538943 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.547552 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: W1204 17:39:24.552584 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod829fda7a_6a96_4007_b2f5_0be03782a764.slice/crio-402c12323607d730312b81d3687659e5ce822fe26ff42fda5f91aa8581f149dc WatchSource:0}: Error finding container 402c12323607d730312b81d3687659e5ce822fe26ff42fda5f91aa8581f149dc: Status 404 returned error can't find the container with id 402c12323607d730312b81d3687659e5ce822fe26ff42fda5f91aa8581f149dc Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.562126 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.580294 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.591693 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.606904 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qp9sx"] Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.609411 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.612544 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.613494 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.614313 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.614637 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.614805 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.614849 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.614882 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.631150 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.641138 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.641948 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.641973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.641983 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.641999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.642009 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.661258 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.679065 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.690108 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.701458 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.715185 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721447 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-log-socket\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721495 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-node-log\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721509 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-ovn-kubernetes\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721533 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-systemd-units\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721546 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-netns\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721568 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-kubelet\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721625 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovn-node-metrics-cert\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721673 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhdsl\" (UniqueName: \"kubernetes.io/projected/067d57d5-c6ec-463d-b1a0-ec5e8c686453-kube-api-access-vhdsl\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721700 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-systemd\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721735 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-ovn\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721784 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-config\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721853 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-var-lib-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721874 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-etc-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721891 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-script-lib\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721912 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-slash\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721926 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-env-overrides\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721951 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-netd\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721972 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-bin\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.721989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.730659 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.744503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.744548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.744559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.744586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.744595 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.747253 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.766431 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.786194 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.800771 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.812728 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823303 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-ovn\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823378 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-config\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823402 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-var-lib-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823424 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-script-lib\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823456 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-ovn\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823484 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-etc-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823492 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-var-lib-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823532 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-slash\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823543 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-etc-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823558 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-env-overrides\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823582 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-slash\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823767 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-netd\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823826 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-bin\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823852 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823868 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-netd\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823937 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823956 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823963 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-bin\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824554 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-config\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.823889 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-openvswitch\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824557 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-env-overrides\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824566 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-script-lib\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824670 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-log-socket\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824724 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-node-log\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824777 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-log-socket\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824851 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-node-log\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824895 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-ovn-kubernetes\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824906 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-ovn-kubernetes\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824953 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-systemd-units\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.824973 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-netns\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.825020 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhdsl\" (UniqueName: \"kubernetes.io/projected/067d57d5-c6ec-463d-b1a0-ec5e8c686453-kube-api-access-vhdsl\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.825030 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-systemd-units\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.825048 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-kubelet\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.825072 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovn-node-metrics-cert\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.825080 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-netns\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.825096 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-systemd\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.825130 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-kubelet\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.825154 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-systemd\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.828467 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovn-node-metrics-cert\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.829828 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:24Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.843883 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhdsl\" (UniqueName: \"kubernetes.io/projected/067d57d5-c6ec-463d-b1a0-ec5e8c686453-kube-api-access-vhdsl\") pod \"ovnkube-node-qp9sx\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.847505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.847535 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.847544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.847558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.847567 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.949773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.949830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.949841 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.949861 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.949872 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:24Z","lastTransitionTime":"2025-12-04T17:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:24 crc kubenswrapper[4733]: I1204 17:39:24.956992 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:24 crc kubenswrapper[4733]: W1204 17:39:24.966686 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod067d57d5_c6ec_463d_b1a0_ec5e8c686453.slice/crio-f4aa35112d541e6a6bc44ab54969cdb9311f1517f34e17ae8ca00f7684347f60 WatchSource:0}: Error finding container f4aa35112d541e6a6bc44ab54969cdb9311f1517f34e17ae8ca00f7684347f60: Status 404 returned error can't find the container with id f4aa35112d541e6a6bc44ab54969cdb9311f1517f34e17ae8ca00f7684347f60 Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.027576 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.027787 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:39:33.027768355 +0000 UTC m=+34.983129411 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.052232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.052267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.052279 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.052297 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.052309 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.131338 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.131600 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.131621 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.131645 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.131527 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.131759 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.131771 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.131829 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:33.131816211 +0000 UTC m=+35.087177257 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.132117 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.132128 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.132135 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.132157 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:33.13215012 +0000 UTC m=+35.087511166 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.132184 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.132204 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:33.132198911 +0000 UTC m=+35.087559957 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.131736 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.132224 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:33.132219302 +0000 UTC m=+35.087580348 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.155711 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.155741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.155749 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.155762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.155771 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.261871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.261920 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.261934 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.261952 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.261970 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.334894 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.335015 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.335074 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.335267 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.335337 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:25 crc kubenswrapper[4733]: E1204 17:39:25.335394 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.364475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.364503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.364511 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.364523 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.364533 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.467124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.467184 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.467203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.467228 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.467247 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.525885 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef97472-7ccc-4d3d-a127-e70bced45051" containerID="435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5" exitCode=0 Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.525979 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" event={"ID":"8ef97472-7ccc-4d3d-a127-e70bced45051","Type":"ContainerDied","Data":"435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.526220 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" event={"ID":"8ef97472-7ccc-4d3d-a127-e70bced45051","Type":"ContainerStarted","Data":"8558088948fb7eb1c7d802bf5a61d5772dfb58586f2ba88eb87c2d292ee27a8d"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.529088 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-472m8" event={"ID":"829fda7a-6a96-4007-b2f5-0be03782a764","Type":"ContainerStarted","Data":"8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.529127 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-472m8" event={"ID":"829fda7a-6a96-4007-b2f5-0be03782a764","Type":"ContainerStarted","Data":"402c12323607d730312b81d3687659e5ce822fe26ff42fda5f91aa8581f149dc"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.535771 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.535839 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.535862 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"7321c9730089f2f1635aa766c37a359fbe2d09e28d3bab564a36d0ee1143218c"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.537109 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"f4aa35112d541e6a6bc44ab54969cdb9311f1517f34e17ae8ca00f7684347f60"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.542919 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6tl82" event={"ID":"21f90af6-f871-4cce-88b4-824c4a15e42c","Type":"ContainerStarted","Data":"322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.542972 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6tl82" event={"ID":"21f90af6-f871-4cce-88b4-824c4a15e42c","Type":"ContainerStarted","Data":"b9cf1fee146fcff454c424cd0dc03784193f214540fc6c49bb29b2c75eb4128f"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.543012 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.556093 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.571414 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.574952 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.574985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.574996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.575014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.575027 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.590498 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.610284 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.635152 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.651228 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.669272 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.677708 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.677741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.677753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.677775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.677813 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.686967 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.698813 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.715283 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.726196 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.738056 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.748667 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.761332 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.781001 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.784277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.784313 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.784322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.784336 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.784344 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.796598 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.812754 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.839154 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.860975 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.877450 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.887242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.887478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.887583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.887691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.887777 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.895048 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.946038 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.965555 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.986114 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.990266 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.990316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.990331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.990351 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:25 crc kubenswrapper[4733]: I1204 17:39:25.990364 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:25Z","lastTransitionTime":"2025-12-04T17:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.001532 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:25Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.014034 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.023919 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.093012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.093044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.093053 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.093066 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.093074 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.195708 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.195765 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.195782 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.195833 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.195850 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.298176 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.298236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.298253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.298296 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.298314 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.400556 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.400603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.400620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.400646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.400664 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.503258 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.503305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.503314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.503329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.503339 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.547687 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef97472-7ccc-4d3d-a127-e70bced45051" containerID="cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45" exitCode=0 Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.547784 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" event={"ID":"8ef97472-7ccc-4d3d-a127-e70bced45051","Type":"ContainerDied","Data":"cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.550376 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18" exitCode=0 Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.550418 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.568119 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.590490 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.608261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.609482 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.609529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.613096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.613604 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.628566 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.652312 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.673149 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.692070 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.716601 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.716633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.716641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.716655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.716664 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.716757 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.735737 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.750415 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.767954 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.782415 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.797193 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.810412 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.819211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.819253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.819265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.819284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.819296 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.832002 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.835450 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-sl9kt"] Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.835875 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.838177 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.838649 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.838952 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.839260 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.863353 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.888956 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.903394 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.916468 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.922339 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.922373 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.922384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.922404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.922416 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:26Z","lastTransitionTime":"2025-12-04T17:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.945472 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.947449 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/660f910c-5b41-457f-aeda-3b6319b67533-host\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.947552 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx5hd\" (UniqueName: \"kubernetes.io/projected/660f910c-5b41-457f-aeda-3b6319b67533-kube-api-access-kx5hd\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.947673 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/660f910c-5b41-457f-aeda-3b6319b67533-serviceca\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.968965 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:26 crc kubenswrapper[4733]: I1204 17:39:26.992068 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:26Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.009610 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.024257 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.026715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.026760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.026773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.026804 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.026816 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.040406 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.049404 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx5hd\" (UniqueName: \"kubernetes.io/projected/660f910c-5b41-457f-aeda-3b6319b67533-kube-api-access-kx5hd\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.049477 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/660f910c-5b41-457f-aeda-3b6319b67533-serviceca\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.049504 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/660f910c-5b41-457f-aeda-3b6319b67533-host\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.049567 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/660f910c-5b41-457f-aeda-3b6319b67533-host\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.051773 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/660f910c-5b41-457f-aeda-3b6319b67533-serviceca\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.059834 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.073247 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.074991 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx5hd\" (UniqueName: \"kubernetes.io/projected/660f910c-5b41-457f-aeda-3b6319b67533-kube-api-access-kx5hd\") pod \"node-ca-sl9kt\" (UID: \"660f910c-5b41-457f-aeda-3b6319b67533\") " pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.093587 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.111041 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.129404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.129456 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.129473 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.129499 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.129519 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.137292 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.151941 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.168321 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.174508 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-sl9kt" Dec 04 17:39:27 crc kubenswrapper[4733]: W1204 17:39:27.186268 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod660f910c_5b41_457f_aeda_3b6319b67533.slice/crio-ec3cc1e1aae4267c8a3b2a316930a0abba2415fe60bab2f0b088e239fc8324bc WatchSource:0}: Error finding container ec3cc1e1aae4267c8a3b2a316930a0abba2415fe60bab2f0b088e239fc8324bc: Status 404 returned error can't find the container with id ec3cc1e1aae4267c8a3b2a316930a0abba2415fe60bab2f0b088e239fc8324bc Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.188941 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.226414 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.232526 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.232586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.232606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.232634 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.232652 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.240531 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.261038 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.280441 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.302504 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.321374 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.335113 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.335120 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:27 crc kubenswrapper[4733]: E1204 17:39:27.335252 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:27 crc kubenswrapper[4733]: E1204 17:39:27.335341 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.335120 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:27 crc kubenswrapper[4733]: E1204 17:39:27.335494 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.336832 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.336903 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.336973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.336998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.337016 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.339520 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.356810 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.377258 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.393237 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.410260 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.442659 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.442702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.442714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.442731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.442747 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.545389 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.545450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.545468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.545493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.545510 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.555575 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-sl9kt" event={"ID":"660f910c-5b41-457f-aeda-3b6319b67533","Type":"ContainerStarted","Data":"6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.555636 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-sl9kt" event={"ID":"660f910c-5b41-457f-aeda-3b6319b67533","Type":"ContainerStarted","Data":"ec3cc1e1aae4267c8a3b2a316930a0abba2415fe60bab2f0b088e239fc8324bc"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.559959 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef97472-7ccc-4d3d-a127-e70bced45051" containerID="2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442" exitCode=0 Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.560055 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" event={"ID":"8ef97472-7ccc-4d3d-a127-e70bced45051","Type":"ContainerDied","Data":"2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.573827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.573875 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.573906 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.573924 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.573941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.573958 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.578302 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.608739 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.625568 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.641959 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.649531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.649602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.649626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.649657 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.649681 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.668709 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.679335 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.696935 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.726358 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.743505 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.752490 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.752515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.752524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.752536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.752546 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.761678 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.773398 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.786509 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.801827 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.814259 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.829740 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.855956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.856010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.856027 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.856052 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.856069 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.857823 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.873207 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.895832 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.932241 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.958452 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.958498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.958514 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.958538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.958555 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:27Z","lastTransitionTime":"2025-12-04T17:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:27 crc kubenswrapper[4733]: I1204 17:39:27.972947 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:27Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.011511 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.056862 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.061180 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.061244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.061269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.061297 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.061322 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.093328 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.135826 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.165691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.165725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.165736 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.165751 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.165771 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.185514 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.213128 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.255877 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.268100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.268161 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.268206 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.268230 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.268248 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.292759 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.345309 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.370542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.370602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.370621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.370646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.370665 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.372161 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.415551 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.461142 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.473403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.473472 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.473496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.473524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.473550 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.511305 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.541111 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.573713 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.576527 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.576568 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.576579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.576596 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.576607 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.581769 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef97472-7ccc-4d3d-a127-e70bced45051" containerID="91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24" exitCode=0 Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.581868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" event={"ID":"8ef97472-7ccc-4d3d-a127-e70bced45051","Type":"ContainerDied","Data":"91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.620997 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.663390 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.679992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.680049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.680067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.680089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.680106 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.693936 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.736840 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.777301 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.783847 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.783882 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.783893 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.783907 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.783917 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.815703 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.853958 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.887015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.887059 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.887078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.887104 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.887123 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.894981 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.929557 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.976901 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:28Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.989842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.989892 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.989910 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.989935 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:28 crc kubenswrapper[4733]: I1204 17:39:28.989953 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:28Z","lastTransitionTime":"2025-12-04T17:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.016205 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.056175 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.108785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.108885 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.108912 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.108947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.108968 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.154814 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.169104 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.188898 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.211197 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.211257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.211277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.211063 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.211302 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.211406 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.250084 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.296596 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.313477 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.313538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.313555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.313579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.313597 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.331960 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.334993 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.335008 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.335095 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.335142 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.335266 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.335377 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.370597 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.414640 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.417497 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.417594 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.417825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.418084 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.418659 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.452436 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.496510 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.521395 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.521434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.521445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.521461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.521472 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.541217 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.570513 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.589304 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef97472-7ccc-4d3d-a127-e70bced45051" containerID="98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066" exitCode=0 Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.589386 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" event={"ID":"8ef97472-7ccc-4d3d-a127-e70bced45051","Type":"ContainerDied","Data":"98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.598411 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.618467 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.624579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.624643 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.624673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.624705 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.624730 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.659946 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.696077 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.728841 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.728887 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.728906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.728929 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.728946 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.736017 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.772900 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.819220 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.831223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.831306 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.831323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.831355 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.831391 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.859850 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.871992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.872021 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.872031 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.872047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.872060 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.889037 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.890770 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.893744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.893775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.893785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.893805 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.893829 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.911921 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.916891 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.916922 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.916933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.916951 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.916964 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.933440 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.937228 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.937253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.937265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.937281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.937293 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.940324 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.953678 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.957423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.957451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.957463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.957477 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.957488 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.976550 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:29 crc kubenswrapper[4733]: E1204 17:39:29.977016 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.978724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.978759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.978771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.978791 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.978803 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:29Z","lastTransitionTime":"2025-12-04T17:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:29 crc kubenswrapper[4733]: I1204 17:39:29.978621 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:29Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.011252 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.055864 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.081339 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.081389 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.081407 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.081430 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.081447 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.097029 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.145322 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.172621 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.184441 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.184497 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.184514 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.184537 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.184555 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.287349 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.287413 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.287432 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.287458 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.287475 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.389825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.389887 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.389904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.389928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.389945 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.493473 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.493542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.493564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.493588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.493604 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.597166 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.597229 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.597249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.597274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.597291 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.607472 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef97472-7ccc-4d3d-a127-e70bced45051" containerID="4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8" exitCode=0 Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.607519 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" event={"ID":"8ef97472-7ccc-4d3d-a127-e70bced45051","Type":"ContainerDied","Data":"4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.646671 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.672791 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.695887 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.705314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.705420 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.705447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.705478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.705502 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.716099 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.748683 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.766592 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.789234 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.808732 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.808782 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.808804 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.808854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.808873 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.812077 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.829723 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.849576 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.864762 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.880349 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.898778 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.911506 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.911576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.911593 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.911617 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.911639 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:30Z","lastTransitionTime":"2025-12-04T17:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.915853 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:30 crc kubenswrapper[4733]: I1204 17:39:30.936938 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:30Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.014676 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.014701 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.014709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.014722 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.014730 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.117512 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.117538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.117545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.117557 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.117573 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.220196 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.220232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.220244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.220259 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.220270 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.322871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.322925 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.322944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.322969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.322992 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.335063 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.335126 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.335191 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:31 crc kubenswrapper[4733]: E1204 17:39:31.335443 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:31 crc kubenswrapper[4733]: E1204 17:39:31.335589 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:31 crc kubenswrapper[4733]: E1204 17:39:31.335677 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.426534 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.426602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.426624 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.426651 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.426670 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.529699 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.529733 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.529741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.529753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.529763 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.618201 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" event={"ID":"8ef97472-7ccc-4d3d-a127-e70bced45051","Type":"ContainerStarted","Data":"fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.632613 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.632672 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.632690 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.632715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.632733 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.643027 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.660530 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.680367 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.702297 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.724545 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.738498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.738583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.738609 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.738641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.738668 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.741266 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.764297 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.781425 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.799720 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.824065 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.838131 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.841558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.841603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.841614 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.841629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.841641 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.864171 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.887308 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.903949 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.920143 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:31Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.945067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.945131 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.945148 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.945174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:31 crc kubenswrapper[4733]: I1204 17:39:31.945193 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:31Z","lastTransitionTime":"2025-12-04T17:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.047938 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.048010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.048036 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.048246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.048268 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.119910 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.136380 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.151696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.151746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.151764 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.151790 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.151839 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.160716 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.187228 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.202444 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.218440 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.232898 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.254609 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.254656 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.254672 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.254693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.254708 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.263348 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.274546 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.301231 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.324982 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.341988 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.355903 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.357646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.357679 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.357690 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.357706 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.357719 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.375471 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.388542 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.406103 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.461394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.461451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.461467 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.461496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.461517 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.564633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.564688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.564702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.564725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.564743 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.641843 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.642267 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.669119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.669662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.669675 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.669699 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.669713 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.676275 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.680086 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.699055 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.720095 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.740126 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.771398 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.773528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.773574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.773591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.773615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.773632 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.788553 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.807931 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.829091 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.846695 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.865751 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.877629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.877654 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.877665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.877683 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.877697 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.881899 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.900677 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.922411 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.938803 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.956352 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.971337 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.979922 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.979984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.979997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.980018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.980031 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:32Z","lastTransitionTime":"2025-12-04T17:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:32 crc kubenswrapper[4733]: I1204 17:39:32.985069 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:32Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.003157 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.013522 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.029960 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.050246 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.069440 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.083465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.083495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.083504 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.083520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.083528 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.086639 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.111121 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.112708 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.112911 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:39:49.112896409 +0000 UTC m=+51.068257455 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.127968 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.145490 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.169104 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.183695 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.186153 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.186202 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.186214 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.186232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.186244 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.214205 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.214256 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.214287 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.214312 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214396 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214458 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214500 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:49.214475176 +0000 UTC m=+51.169836232 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214539 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214579 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214580 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214639 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214654 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214600 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214543 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:49.214522618 +0000 UTC m=+51.169883774 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214784 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:49.214770494 +0000 UTC m=+51.170131660 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.214816 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 17:39:49.214808166 +0000 UTC m=+51.170169342 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.215996 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.244833 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.288380 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.288437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.288446 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.288459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.288468 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.335495 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.335559 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.335507 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.335714 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.335959 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:33 crc kubenswrapper[4733]: E1204 17:39:33.336087 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.391638 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.391712 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.391729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.391757 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.391777 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.494740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.494831 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.494849 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.494876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.494900 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.597941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.598015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.598038 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.598067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.598091 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.645224 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.646305 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.681518 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.701278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.701339 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.701358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.701385 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.701403 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.705693 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.724434 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.739952 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.757079 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.769612 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.791376 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.804145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.804243 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.804261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.804313 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.804331 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.806285 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.826205 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.849167 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.870803 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.884602 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.904687 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.906906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.906960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.906983 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.907016 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.907038 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:33Z","lastTransitionTime":"2025-12-04T17:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.934605 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.948094 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:33 crc kubenswrapper[4733]: I1204 17:39:33.967668 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:33Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.010179 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.010224 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.010240 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.010262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.010278 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.112635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.112679 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.112692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.112712 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.112725 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.215580 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.215635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.215656 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.215684 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.215704 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.319571 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.319974 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.320257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.320453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.320637 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.424365 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.424409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.424426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.424449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.424465 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.531284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.531364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.531389 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.531422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.531446 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.634031 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.634075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.634086 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.634105 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.634116 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.651442 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/0.log" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.655517 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00" exitCode=1 Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.655581 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.656606 4733 scope.go:117] "RemoveContainer" containerID="aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.687799 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:34Z\\\",\\\"message\\\":\\\" 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 17:39:33.902259 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 17:39:33.902295 6035 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 17:39:33.902316 6035 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 17:39:33.902355 6035 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 17:39:33.902369 6035 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 17:39:33.902399 6035 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 17:39:33.903312 6035 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 17:39:33.903320 6035 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 17:39:33.903328 6035 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 17:39:33.903346 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 17:39:33.903535 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 17:39:33.903572 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 17:39:33.903609 6035 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 17:39:33.903629 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 17:39:33.903675 6035 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.704376 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.735786 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.737454 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.737520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.737542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.737570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.737592 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.759281 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.776051 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.798125 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.814052 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.833335 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.840279 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.840345 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.840367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.840399 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.840421 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.849767 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.867349 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.880777 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.892946 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.912402 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.930410 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.942287 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.942325 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.942337 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.942352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.942362 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:34Z","lastTransitionTime":"2025-12-04T17:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:34 crc kubenswrapper[4733]: I1204 17:39:34.948663 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:34Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.044642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.044674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.044682 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.044697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.044706 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.147328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.147375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.147386 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.147402 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.147413 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.250665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.250698 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.250708 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.250721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.250729 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.334475 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.334503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.334572 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:35 crc kubenswrapper[4733]: E1204 17:39:35.334600 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:35 crc kubenswrapper[4733]: E1204 17:39:35.334680 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:35 crc kubenswrapper[4733]: E1204 17:39:35.334862 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.356089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.356165 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.356184 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.356213 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.356232 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.459200 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.459276 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.459301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.459362 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.459383 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.562116 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.562152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.562162 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.562178 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.562189 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.662375 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/0.log" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.664753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.664844 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.664866 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.664888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.664906 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.667014 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.667448 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.692374 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.715400 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.738066 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.757065 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.767158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.767210 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.767225 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.767245 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.767259 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.780476 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.791206 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.802052 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.823342 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.850867 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.869692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.869758 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.869778 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.869860 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.869882 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.874536 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.895910 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.918379 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.950992 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:34Z\\\",\\\"message\\\":\\\" 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 17:39:33.902259 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 17:39:33.902295 6035 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 17:39:33.902316 6035 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 17:39:33.902355 6035 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 17:39:33.902369 6035 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 17:39:33.902399 6035 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 17:39:33.903312 6035 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 17:39:33.903320 6035 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 17:39:33.903328 6035 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 17:39:33.903346 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 17:39:33.903535 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 17:39:33.903572 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 17:39:33.903609 6035 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 17:39:33.903629 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 17:39:33.903675 6035 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.965545 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.972535 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.972590 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.972611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.972640 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.972660 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:35Z","lastTransitionTime":"2025-12-04T17:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:35 crc kubenswrapper[4733]: I1204 17:39:35.998436 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.019463 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.075867 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.075911 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.075923 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.075940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.075953 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.177748 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.177825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.177839 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.177857 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.177868 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.280413 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.280479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.280495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.280521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.280540 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.383947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.384014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.384031 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.384055 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.384073 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.487777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.487888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.487917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.487943 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.487961 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.590395 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.590446 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.590456 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.590470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.590481 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.672879 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/1.log" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.674343 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/0.log" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.678669 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a" exitCode=1 Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.678728 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.678834 4733 scope.go:117] "RemoveContainer" containerID="aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.679952 4733 scope.go:117] "RemoveContainer" containerID="5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a" Dec 04 17:39:36 crc kubenswrapper[4733]: E1204 17:39:36.680193 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.697845 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.697909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.697926 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.697951 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.697968 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.712253 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.735380 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.758191 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.777908 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.793279 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.800374 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.800417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.800455 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.800477 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.800491 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.813451 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.848405 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:34Z\\\",\\\"message\\\":\\\" 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 17:39:33.902259 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 17:39:33.902295 6035 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 17:39:33.902316 6035 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 17:39:33.902355 6035 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 17:39:33.902369 6035 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 17:39:33.902399 6035 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 17:39:33.903312 6035 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 17:39:33.903320 6035 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 17:39:33.903328 6035 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 17:39:33.903346 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 17:39:33.903535 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 17:39:33.903572 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 17:39:33.903609 6035 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 17:39:33.903629 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 17:39:33.903675 6035 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.865063 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.897493 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.902971 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.903016 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.903033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.903056 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.903074 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:36Z","lastTransitionTime":"2025-12-04T17:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.920625 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.937086 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.953938 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.975208 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:36 crc kubenswrapper[4733]: I1204 17:39:36.990836 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:36Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.006076 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.006154 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.006173 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.006200 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.006218 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.009760 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.109058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.109138 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.109163 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.109203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.109228 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.116753 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2"] Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.117613 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.121061 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.121236 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.144047 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.161968 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/00bca008-075a-4027-944f-b3c9cad33814-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.162042 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nwhb\" (UniqueName: \"kubernetes.io/projected/00bca008-075a-4027-944f-b3c9cad33814-kube-api-access-6nwhb\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.162078 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00bca008-075a-4027-944f-b3c9cad33814-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.162127 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/00bca008-075a-4027-944f-b3c9cad33814-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.163971 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.184075 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.201744 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.212122 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.212169 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.212186 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.212207 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.212225 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.225382 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.243880 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.263366 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/00bca008-075a-4027-944f-b3c9cad33814-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.263679 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/00bca008-075a-4027-944f-b3c9cad33814-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.263935 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nwhb\" (UniqueName: \"kubernetes.io/projected/00bca008-075a-4027-944f-b3c9cad33814-kube-api-access-6nwhb\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.264119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00bca008-075a-4027-944f-b3c9cad33814-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.264531 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/00bca008-075a-4027-944f-b3c9cad33814-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.264582 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/00bca008-075a-4027-944f-b3c9cad33814-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.265401 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.274159 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00bca008-075a-4027-944f-b3c9cad33814-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.288124 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nwhb\" (UniqueName: \"kubernetes.io/projected/00bca008-075a-4027-944f-b3c9cad33814-kube-api-access-6nwhb\") pod \"ovnkube-control-plane-749d76644c-rv2n2\" (UID: \"00bca008-075a-4027-944f-b3c9cad33814\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.292011 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.312284 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.316400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.316479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.316505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.316535 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.316559 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.331213 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.335005 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.335086 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:37 crc kubenswrapper[4733]: E1204 17:39:37.335368 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.335109 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:37 crc kubenswrapper[4733]: E1204 17:39:37.335454 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:37 crc kubenswrapper[4733]: E1204 17:39:37.335386 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.351438 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.375583 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec8b0b335c298b978d5c1ebce210c4c58da8eed128b7e80dc24699147a94f00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:34Z\\\",\\\"message\\\":\\\" 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 17:39:33.902259 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 17:39:33.902295 6035 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 17:39:33.902316 6035 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 17:39:33.902355 6035 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 17:39:33.902369 6035 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 17:39:33.902399 6035 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 17:39:33.903312 6035 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 17:39:33.903320 6035 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1204 17:39:33.903328 6035 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 17:39:33.903346 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 17:39:33.903535 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 17:39:33.903572 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 17:39:33.903609 6035 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 17:39:33.903629 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 17:39:33.903675 6035 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.390831 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.409104 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.422618 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.422702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.422729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.422764 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.422861 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.433628 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.446510 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.477298 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.526763 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.526856 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.526874 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.526899 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.526918 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.630255 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.630314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.630334 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.630358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.630377 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.684033 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" event={"ID":"00bca008-075a-4027-944f-b3c9cad33814","Type":"ContainerStarted","Data":"fb10da302e0d2a4bafce20394bb0b8d22b42384f337e2e198c13e22f08149194"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.686752 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/1.log" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.692068 4733 scope.go:117] "RemoveContainer" containerID="5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a" Dec 04 17:39:37 crc kubenswrapper[4733]: E1204 17:39:37.692331 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.712752 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.735513 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.735545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.735556 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.735574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.735585 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.736608 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.750969 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.768720 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.793072 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.817141 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.833055 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.839389 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.839451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.839469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.839492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.839508 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.846366 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.864025 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.876483 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.899056 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.917471 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.930101 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.942453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.942503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.942521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.942542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.942558 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:37Z","lastTransitionTime":"2025-12-04T17:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.945242 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.959235 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:37 crc kubenswrapper[4733]: I1204 17:39:37.977470 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:37Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.045326 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.045372 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.045384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.045402 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.045415 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.147519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.147575 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.147591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.147615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.147632 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.249912 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.250423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.250450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.250481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.250505 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.279485 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-mlwj8"] Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.280255 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:38 crc kubenswrapper[4733]: E1204 17:39:38.280363 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.300296 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.312625 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.328319 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.348733 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.354221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.354331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.354417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.354493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.354566 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.364327 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.376649 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.376861 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kp2k\" (UniqueName: \"kubernetes.io/projected/79d2d07f-2f47-4970-b6e5-7d25b02e639e-kube-api-access-7kp2k\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.379573 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.395474 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.411948 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.439846 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.454586 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.456533 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.456583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.456601 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.456625 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.456642 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.478054 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kp2k\" (UniqueName: \"kubernetes.io/projected/79d2d07f-2f47-4970-b6e5-7d25b02e639e-kube-api-access-7kp2k\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.478191 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:38 crc kubenswrapper[4733]: E1204 17:39:38.478333 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:38 crc kubenswrapper[4733]: E1204 17:39:38.478433 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs podName:79d2d07f-2f47-4970-b6e5-7d25b02e639e nodeName:}" failed. No retries permitted until 2025-12-04 17:39:38.978408312 +0000 UTC m=+40.933769388 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs") pod "network-metrics-daemon-mlwj8" (UID: "79d2d07f-2f47-4970-b6e5-7d25b02e639e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.488019 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.498287 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kp2k\" (UniqueName: \"kubernetes.io/projected/79d2d07f-2f47-4970-b6e5-7d25b02e639e-kube-api-access-7kp2k\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.507032 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.529578 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.549666 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.559041 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.559151 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.559183 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.559210 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.559227 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.569962 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.585342 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.602651 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.637731 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.660549 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.662090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.662328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.662455 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.662720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.662877 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.678651 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.701135 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" event={"ID":"00bca008-075a-4027-944f-b3c9cad33814","Type":"ContainerStarted","Data":"05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.701139 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.701222 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" event={"ID":"00bca008-075a-4027-944f-b3c9cad33814","Type":"ContainerStarted","Data":"0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.728299 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.742265 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.753541 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.765566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.765621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.765635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.765683 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.765697 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.772133 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.785773 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.803484 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.823292 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.838552 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.854712 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.868169 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.868219 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.868237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.868263 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.868281 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.875658 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.896751 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.920950 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.932676 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.946475 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.961303 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.970421 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.970472 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.970483 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.970497 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.970508 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:38Z","lastTransitionTime":"2025-12-04T17:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.977465 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.983357 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:38 crc kubenswrapper[4733]: E1204 17:39:38.983541 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:38 crc kubenswrapper[4733]: E1204 17:39:38.983609 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs podName:79d2d07f-2f47-4970-b6e5-7d25b02e639e nodeName:}" failed. No retries permitted until 2025-12-04 17:39:39.983590892 +0000 UTC m=+41.938951958 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs") pod "network-metrics-daemon-mlwj8" (UID: "79d2d07f-2f47-4970-b6e5-7d25b02e639e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:38 crc kubenswrapper[4733]: I1204 17:39:38.993264 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:38Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.006233 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.022200 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.036599 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.050495 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.073621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.073683 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.073697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.073709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.073720 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.077782 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.092894 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.108372 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.131329 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.146081 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.161227 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.176580 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.176683 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.176764 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.176782 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.176842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.176862 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.191921 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.207133 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:39Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.279953 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.280036 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.280061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.280089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.280108 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.335346 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.335410 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:39 crc kubenswrapper[4733]: E1204 17:39:39.335561 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.344118 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:39 crc kubenswrapper[4733]: E1204 17:39:39.344267 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:39 crc kubenswrapper[4733]: E1204 17:39:39.344455 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.381972 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.382023 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.382038 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.382057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.382069 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.485232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.485280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.485290 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.485305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.485314 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.587377 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.587442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.587464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.587493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.587517 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.690146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.690206 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.690217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.690233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.690244 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.793271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.793444 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.793466 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.793488 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.793560 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.897174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.897255 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.897272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.897297 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.897315 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:39Z","lastTransitionTime":"2025-12-04T17:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:39 crc kubenswrapper[4733]: I1204 17:39:39.993441 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:39 crc kubenswrapper[4733]: E1204 17:39:39.993625 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:39 crc kubenswrapper[4733]: E1204 17:39:39.993694 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs podName:79d2d07f-2f47-4970-b6e5-7d25b02e639e nodeName:}" failed. No retries permitted until 2025-12-04 17:39:41.993674885 +0000 UTC m=+43.949035931 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs") pod "network-metrics-daemon-mlwj8" (UID: "79d2d07f-2f47-4970-b6e5-7d25b02e639e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.000398 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.000443 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.000464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.000491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.000511 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.103467 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.103513 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.103529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.103551 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.103608 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.206360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.206433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.206458 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.206487 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.206509 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.293478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.293529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.293546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.293568 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.293585 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: E1204 17:39:40.316109 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:40Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.322332 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.322426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.322455 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.322489 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.322514 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.335306 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:40 crc kubenswrapper[4733]: E1204 17:39:40.335538 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:40 crc kubenswrapper[4733]: E1204 17:39:40.344933 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:40Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.352984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.353092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.353115 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.353160 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.353181 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: E1204 17:39:40.375174 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:40Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.381182 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.381253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.381278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.381303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.381322 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: E1204 17:39:40.403009 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:40Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.407415 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.407485 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.407503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.407528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.407552 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: E1204 17:39:40.428563 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:40Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:40 crc kubenswrapper[4733]: E1204 17:39:40.428833 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.430967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.431018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.431035 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.431059 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.431077 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.533622 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.533716 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.533735 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.533761 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.533778 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.637086 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.637138 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.637156 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.637178 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.637196 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.740589 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.740629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.740663 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.740678 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.740688 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.843541 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.843588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.843600 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.843621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.843632 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.946124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.946192 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.946216 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.946248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:40 crc kubenswrapper[4733]: I1204 17:39:40.946267 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:40Z","lastTransitionTime":"2025-12-04T17:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.049318 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.049385 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.049402 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.049425 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.049444 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.152629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.152696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.152711 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.152736 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.152757 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.256766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.256878 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.256902 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.256934 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.256965 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.334967 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.335168 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:41 crc kubenswrapper[4733]: E1204 17:39:41.335238 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.335428 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:41 crc kubenswrapper[4733]: E1204 17:39:41.335614 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:41 crc kubenswrapper[4733]: E1204 17:39:41.335707 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.360187 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.360254 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.360271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.360298 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.360317 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.462492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.462546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.462563 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.462587 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.462604 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.565910 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.565973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.565991 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.566014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.566030 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.669189 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.669246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.669262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.669289 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.669312 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.772561 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.772625 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.772644 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.772675 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.772693 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.876288 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.876349 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.876368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.876392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.876407 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.979301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.979403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.979426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.979459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:41 crc kubenswrapper[4733]: I1204 17:39:41.979482 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:41Z","lastTransitionTime":"2025-12-04T17:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.025239 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:42 crc kubenswrapper[4733]: E1204 17:39:42.025427 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:42 crc kubenswrapper[4733]: E1204 17:39:42.025530 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs podName:79d2d07f-2f47-4970-b6e5-7d25b02e639e nodeName:}" failed. No retries permitted until 2025-12-04 17:39:46.025499826 +0000 UTC m=+47.980860912 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs") pod "network-metrics-daemon-mlwj8" (UID: "79d2d07f-2f47-4970-b6e5-7d25b02e639e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.082892 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.082944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.082961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.082987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.083004 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.187233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.187320 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.187333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.187358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.187374 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.290936 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.291006 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.291027 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.291057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.291079 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.335384 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:42 crc kubenswrapper[4733]: E1204 17:39:42.335638 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.394187 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.394222 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.394233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.394250 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.394262 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.497949 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.498003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.498071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.498098 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.498118 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.601419 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.601473 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.601486 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.601506 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.601521 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.704505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.704570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.704586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.704614 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.704637 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.807947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.808013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.808030 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.808055 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.808072 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.920574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.920637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.920654 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.920677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:42 crc kubenswrapper[4733]: I1204 17:39:42.920696 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:42Z","lastTransitionTime":"2025-12-04T17:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.024331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.024420 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.024440 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.024475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.024494 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.128281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.128336 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.128355 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.128378 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.128397 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.231959 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.232014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.232032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.232055 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.232072 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.334317 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.334416 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:43 crc kubenswrapper[4733]: E1204 17:39:43.334501 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.334539 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:43 crc kubenswrapper[4733]: E1204 17:39:43.334726 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:43 crc kubenswrapper[4733]: E1204 17:39:43.334961 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.335011 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.335050 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.335070 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.335092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.335109 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.437786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.437887 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.437938 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.437962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.437982 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.541863 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.541955 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.541973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.542031 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.542048 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.645056 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.645119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.645137 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.645162 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.645181 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.747973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.748069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.748087 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.748113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.748132 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.851384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.851462 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.851481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.851509 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.851535 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.955372 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.955436 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.955453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.955481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:43 crc kubenswrapper[4733]: I1204 17:39:43.955505 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:43Z","lastTransitionTime":"2025-12-04T17:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.058987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.059053 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.059070 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.059096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.059118 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.162401 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.162543 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.162565 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.162592 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.162614 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.265979 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.266062 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.266079 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.266111 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.266131 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.335195 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:44 crc kubenswrapper[4733]: E1204 17:39:44.335421 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.369868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.369969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.369992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.370065 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.370085 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.472892 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.472926 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.472937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.472952 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.472960 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.576479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.576553 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.576570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.576594 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.576612 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.679691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.679753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.679766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.679784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.679822 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.782227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.782272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.782287 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.782311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.782347 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.885578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.885645 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.885667 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.885694 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.885717 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.989221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.989286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.989306 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.989334 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:44 crc kubenswrapper[4733]: I1204 17:39:44.989353 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:44Z","lastTransitionTime":"2025-12-04T17:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.092174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.092219 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.092236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.092259 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.092275 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.196109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.196169 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.196193 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.196219 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.196236 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.301172 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.301245 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.301271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.301303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.301331 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.334851 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.334855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:45 crc kubenswrapper[4733]: E1204 17:39:45.335111 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.335109 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:45 crc kubenswrapper[4733]: E1204 17:39:45.335742 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:45 crc kubenswrapper[4733]: E1204 17:39:45.335896 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.406042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.406140 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.406158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.406180 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.406198 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.511536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.511583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.511594 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.511615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.511628 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.615648 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.615715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.615724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.615748 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.615760 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.719103 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.719152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.719169 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.719194 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.719212 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.822281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.822350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.822371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.822397 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.822415 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.925377 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.925434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.925452 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.925475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:45 crc kubenswrapper[4733]: I1204 17:39:45.925493 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:45Z","lastTransitionTime":"2025-12-04T17:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.027480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.027531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.027543 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.027563 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.027576 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.068354 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:46 crc kubenswrapper[4733]: E1204 17:39:46.068510 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:46 crc kubenswrapper[4733]: E1204 17:39:46.068564 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs podName:79d2d07f-2f47-4970-b6e5-7d25b02e639e nodeName:}" failed. No retries permitted until 2025-12-04 17:39:54.068549031 +0000 UTC m=+56.023910077 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs") pod "network-metrics-daemon-mlwj8" (UID: "79d2d07f-2f47-4970-b6e5-7d25b02e639e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.129981 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.130017 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.130031 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.130049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.130064 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.232955 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.233006 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.233021 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.233044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.233065 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.334517 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:46 crc kubenswrapper[4733]: E1204 17:39:46.334714 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.336448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.336518 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.336538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.336564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.336581 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.439368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.439416 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.439427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.439444 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.439456 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.542406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.542491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.542515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.542545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.542569 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.645603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.645651 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.645659 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.645674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.645684 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.748787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.748904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.748942 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.748974 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.748998 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.852393 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.852451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.852472 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.852502 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.852525 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.954849 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.954899 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.954915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.954931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:46 crc kubenswrapper[4733]: I1204 17:39:46.954942 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:46Z","lastTransitionTime":"2025-12-04T17:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.057359 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.057419 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.057432 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.057450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.057462 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.159674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.159744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.159771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.159837 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.159864 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.263120 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.263173 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.263190 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.263215 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.263233 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.334957 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.335042 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.335084 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:47 crc kubenswrapper[4733]: E1204 17:39:47.335170 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:47 crc kubenswrapper[4733]: E1204 17:39:47.335276 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:47 crc kubenswrapper[4733]: E1204 17:39:47.335408 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.366504 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.366565 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.366588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.366616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.366644 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.470196 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.470291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.470310 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.470352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.470370 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.574558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.574631 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.574653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.574680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.574701 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.677603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.677646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.677656 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.677673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.677687 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.780405 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.780470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.780488 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.780513 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.780532 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.883246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.883312 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.883329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.883353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.883372 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.986478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.986535 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.986558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.986588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:47 crc kubenswrapper[4733]: I1204 17:39:47.986613 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:47Z","lastTransitionTime":"2025-12-04T17:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.089849 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.089924 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.089944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.089973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.089996 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.193142 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.193204 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.193223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.193253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.193275 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.296546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.296616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.296638 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.296668 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.296692 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.335253 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:48 crc kubenswrapper[4733]: E1204 17:39:48.335459 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.337216 4733 scope.go:117] "RemoveContainer" containerID="5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.369350 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.392380 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.401407 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.401463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.401481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.401504 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.401521 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.413275 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.432340 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.464069 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.483611 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.503082 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.505028 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.505094 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.505121 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.505151 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.505172 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.521181 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.538460 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.551046 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.564213 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.575709 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.593482 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.607847 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.607880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.607895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.607913 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.607927 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.613938 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.628158 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.648595 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.660769 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.710163 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.710201 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.710210 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.710224 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.710237 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.749586 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/1.log" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.752439 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.753035 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.770863 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.788572 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.805887 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.812620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.812659 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.812671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.812689 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.812703 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.818238 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.838776 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.855249 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.872030 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.888891 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.905108 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.915048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.915087 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.915096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.915108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.915118 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:48Z","lastTransitionTime":"2025-12-04T17:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.917258 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.928617 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.953276 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.968358 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.986062 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:48 crc kubenswrapper[4733]: I1204 17:39:48.999759 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:48Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.017005 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.017081 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.017091 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.017104 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.017112 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.019587 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.027942 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.119155 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.119191 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.119202 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.119218 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.119230 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.203279 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.203475 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:40:21.203447381 +0000 UTC m=+83.158808437 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.227760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.228336 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.228350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.228371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.228384 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.305118 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.305165 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.305193 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.305229 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305352 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305391 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305407 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305457 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305480 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305422 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:40:21.30540245 +0000 UTC m=+83.260763496 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305537 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305584 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:40:21.305556014 +0000 UTC m=+83.260917100 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305633 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 17:40:21.305616276 +0000 UTC m=+83.260977362 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305611 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305664 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.305751 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 17:40:21.305724269 +0000 UTC m=+83.261085345 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.331952 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.332015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.332034 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.332063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.332082 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.335196 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.335286 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.335316 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.335507 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.335571 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.335860 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.435117 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.435211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.435233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.435259 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.435309 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.538930 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.539014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.539038 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.539074 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.539097 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.642650 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.642738 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.642760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.642843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.642863 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.746199 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.746265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.746283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.746308 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.746326 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.760020 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/2.log" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.761344 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/1.log" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.766173 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4" exitCode=1 Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.766229 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.766286 4733 scope.go:117] "RemoveContainer" containerID="5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.767573 4733 scope.go:117] "RemoveContainer" containerID="4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4" Dec 04 17:39:49 crc kubenswrapper[4733]: E1204 17:39:49.767928 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.787996 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.809113 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.825653 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.842838 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.848621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.848662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.848675 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.848694 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.848708 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.872199 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.892829 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.911789 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.932014 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.952013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.952076 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.952094 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.952120 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.952138 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:49Z","lastTransitionTime":"2025-12-04T17:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.964860 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2e69a83f37bb627863a6063991b3d2a4f91c7287d070e5fbfae3c7249ee35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:35Z\\\",\\\"message\\\":\\\"1f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:39:35.534626 6156 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1204 17:39:35.534519 6156 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:35Z is after 2025-08-24T17:21:41Z]\\\\nI1204 17:39:35.534912 6156 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 17:39:35.535529 61\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:49 crc kubenswrapper[4733]: I1204 17:39:49.982946 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:49Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.003256 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.025597 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.044522 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.055315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.055394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.055414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.055441 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.055461 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.063366 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.078558 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.094472 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.114508 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.159045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.159111 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.159129 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.159153 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.159170 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.262885 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.262970 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.262999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.263033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.263055 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.334609 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:50 crc kubenswrapper[4733]: E1204 17:39:50.334846 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.365562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.365625 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.365641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.365664 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.365682 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.468453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.468524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.468548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.468577 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.468599 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.572124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.572232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.572258 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.572290 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.572311 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.664480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.664587 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.664611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.664640 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.664659 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: E1204 17:39:50.686624 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.691466 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.691555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.691574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.691595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.691611 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: E1204 17:39:50.710537 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.715071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.715103 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.715113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.715128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.715139 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: E1204 17:39:50.739589 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.744774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.744866 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.744883 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.744911 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.744931 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: E1204 17:39:50.761031 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.765231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.765280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.765307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.765321 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.765331 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.772585 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/2.log" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.778455 4733 scope.go:117] "RemoveContainer" containerID="4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4" Dec 04 17:39:50 crc kubenswrapper[4733]: E1204 17:39:50.778600 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:39:50 crc kubenswrapper[4733]: E1204 17:39:50.782958 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: E1204 17:39:50.783059 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.785057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.785082 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.785091 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.785103 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.785111 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.795194 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.809016 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.826155 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.839979 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.853670 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.877685 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.887707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.887784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.887813 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.887830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.887842 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.895524 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.912866 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.930103 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.960632 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.977677 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.990405 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.990457 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.990474 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.990500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.990519 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:50Z","lastTransitionTime":"2025-12-04T17:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:50 crc kubenswrapper[4733]: I1204 17:39:50.995851 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:50Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.016056 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:51Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.038131 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:51Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.055016 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:51Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.074407 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:51Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.091445 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:51Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.093400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.093501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.093520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.093585 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.093605 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.197266 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.197330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.197346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.197371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.197389 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.300163 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.300223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.300241 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.300267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.300284 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.335501 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.335540 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:51 crc kubenswrapper[4733]: E1204 17:39:51.335691 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.335729 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:51 crc kubenswrapper[4733]: E1204 17:39:51.335901 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:51 crc kubenswrapper[4733]: E1204 17:39:51.336172 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.403254 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.403311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.403330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.403352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.403368 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.505422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.505449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.505456 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.505468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.505477 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.609460 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.609525 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.609548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.609576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.609593 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.713042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.713145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.713166 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.713230 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.713251 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.815734 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.815791 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.815843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.815868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.815895 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.918914 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.918957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.918975 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.918997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:51 crc kubenswrapper[4733]: I1204 17:39:51.919013 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:51Z","lastTransitionTime":"2025-12-04T17:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.022791 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.022875 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.022891 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.022914 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.022936 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.126406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.126475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.126493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.126517 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.126536 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.229216 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.229276 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.229293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.229316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.229333 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.332487 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.332564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.332584 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.332611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.332628 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.334971 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:52 crc kubenswrapper[4733]: E1204 17:39:52.335170 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.435912 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.435975 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.435992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.436018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.436038 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.538989 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.539045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.539062 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.539088 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.539106 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.642329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.642391 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.642410 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.642434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.642452 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.745157 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.745216 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.745233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.745256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.745272 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.847890 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.847957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.847969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.847985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.847998 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.951438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.951515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.951531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.951555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:52 crc kubenswrapper[4733]: I1204 17:39:52.951572 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:52Z","lastTransitionTime":"2025-12-04T17:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.054951 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.055010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.055026 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.055051 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.055069 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.158169 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.158228 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.158244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.158271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.158292 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.261292 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.261350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.261366 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.261384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.261396 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.290004 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.305707 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.325034 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.334378 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.334490 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.334533 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:53 crc kubenswrapper[4733]: E1204 17:39:53.334731 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:53 crc kubenswrapper[4733]: E1204 17:39:53.334865 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:53 crc kubenswrapper[4733]: E1204 17:39:53.335152 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.349308 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.364574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.364621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.364639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.364662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.364679 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.369888 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.389255 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.422306 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.440973 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.459901 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.467747 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.467830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.467856 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.467880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.467898 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.480832 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.501692 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.520404 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.540870 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.558738 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.570986 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.571050 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.571067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.571093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.571119 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.577758 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.593050 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.612213 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.635176 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.650675 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:53Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.675223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.675268 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.675283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.675307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.675324 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.778177 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.778220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.778236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.778257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.778276 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.881109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.881145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.881156 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.881171 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.881182 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.985064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.985132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.985155 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.985185 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:53 crc kubenswrapper[4733]: I1204 17:39:53.985206 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:53Z","lastTransitionTime":"2025-12-04T17:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.088341 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.088460 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.088481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.088507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.088542 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.163310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:54 crc kubenswrapper[4733]: E1204 17:39:54.163462 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:54 crc kubenswrapper[4733]: E1204 17:39:54.163524 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs podName:79d2d07f-2f47-4970-b6e5-7d25b02e639e nodeName:}" failed. No retries permitted until 2025-12-04 17:40:10.163506851 +0000 UTC m=+72.118867897 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs") pod "network-metrics-daemon-mlwj8" (UID: "79d2d07f-2f47-4970-b6e5-7d25b02e639e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.191302 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.191375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.191388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.191406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.191445 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.293388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.293536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.293818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.293848 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.293865 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.338278 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:54 crc kubenswrapper[4733]: E1204 17:39:54.338422 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.397096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.397171 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.397194 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.397225 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.397249 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.499939 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.500440 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.500665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.500696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.500715 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.604158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.604201 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.604220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.604244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.604260 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.707427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.707491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.707508 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.707537 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.707555 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.811078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.811158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.811182 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.811212 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.811234 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.914577 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.914664 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.914691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.914721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:54 crc kubenswrapper[4733]: I1204 17:39:54.914744 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:54Z","lastTransitionTime":"2025-12-04T17:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.017867 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.017928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.017946 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.017971 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.017991 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.120856 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.120954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.120977 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.121008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.121031 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.224462 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.224539 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.224562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.224592 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.224615 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.327673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.327735 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.327752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.327776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.327827 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.335298 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.335364 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.335401 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:55 crc kubenswrapper[4733]: E1204 17:39:55.335503 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:55 crc kubenswrapper[4733]: E1204 17:39:55.335610 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:55 crc kubenswrapper[4733]: E1204 17:39:55.335715 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.433095 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.433171 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.433190 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.433216 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.433235 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.536231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.536286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.536303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.536323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.536336 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.638769 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.638918 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.638937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.638959 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.638974 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.742265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.742320 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.742331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.742350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.742364 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.845327 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.845375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.845386 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.845403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.845415 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.948896 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.948971 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.948994 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.949024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:55 crc kubenswrapper[4733]: I1204 17:39:55.949047 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:55Z","lastTransitionTime":"2025-12-04T17:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.052378 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.052444 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.052467 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.052496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.052519 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.155982 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.156047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.156066 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.156092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.156111 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.259180 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.259232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.259243 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.259260 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.259273 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.334970 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:56 crc kubenswrapper[4733]: E1204 17:39:56.335175 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.361933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.361998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.362015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.362037 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.362054 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.464161 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.464205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.464217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.464234 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.464245 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.567543 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.567626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.567677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.567746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.567769 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.670109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.670140 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.670152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.670168 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.670178 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.772419 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.772466 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.772478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.772500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.772511 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.875600 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.876093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.876114 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.876155 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.876173 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.979242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.979305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.979324 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.979352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:56 crc kubenswrapper[4733]: I1204 17:39:56.979382 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:56Z","lastTransitionTime":"2025-12-04T17:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.082404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.082471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.082493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.082521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.082542 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.185925 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.185984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.186001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.186026 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.186044 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.288993 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.289060 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.289077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.289101 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.289118 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.335200 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.335229 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.335280 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:57 crc kubenswrapper[4733]: E1204 17:39:57.335387 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:57 crc kubenswrapper[4733]: E1204 17:39:57.335619 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:57 crc kubenswrapper[4733]: E1204 17:39:57.335675 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.391692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.391758 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.391782 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.391858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.391882 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.495441 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.495500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.495517 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.495545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.495562 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.598510 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.598589 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.598614 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.598644 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.598664 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.702069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.702133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.702150 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.702174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.702192 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.806941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.806992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.807007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.807029 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.807046 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.909419 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.909490 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.909510 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.909554 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:57 crc kubenswrapper[4733]: I1204 17:39:57.909586 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:57Z","lastTransitionTime":"2025-12-04T17:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.012888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.012954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.012972 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.012997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.013016 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.116379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.116449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.116470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.116500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.116523 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.219876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.219946 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.219963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.219988 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.220006 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.322706 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.322825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.322845 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.322871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.322888 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.334545 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:39:58 crc kubenswrapper[4733]: E1204 17:39:58.334778 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.353671 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.374304 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.393039 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45d07789-93de-4a10-85ee-5af716a86700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.412057 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.426588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.426844 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.427001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.427137 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.427280 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.439020 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.462030 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.492975 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.507915 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.522276 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.530545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.530632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.530651 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.530672 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.530718 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.541418 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.560356 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.581746 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.601525 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.622602 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.633780 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.634017 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.634126 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.634215 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.634295 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.639864 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.660121 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.681558 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.693643 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:39:58Z is after 2025-08-24T17:21:41Z" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.736311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.736370 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.736387 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.736415 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.736433 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.839003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.839075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.839092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.839116 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.839134 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.941985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.942054 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.942075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.942106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:58 crc kubenswrapper[4733]: I1204 17:39:58.942127 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:58Z","lastTransitionTime":"2025-12-04T17:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.045531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.045600 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.045624 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.045655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.045677 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.148728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.148764 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.148772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.148784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.148813 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.251606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.251667 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.251684 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.251709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.251726 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.335287 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.335368 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:39:59 crc kubenswrapper[4733]: E1204 17:39:59.335454 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.335380 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:39:59 crc kubenswrapper[4733]: E1204 17:39:59.335556 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:39:59 crc kubenswrapper[4733]: E1204 17:39:59.335696 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.355167 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.355223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.355241 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.355262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.355280 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.458149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.458203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.458220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.458242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.458258 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.560715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.560783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.560840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.560865 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.560895 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.663723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.663840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.663869 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.663899 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.663921 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.766978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.767069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.767094 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.767129 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.767153 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.870730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.870834 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.870855 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.870880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.870902 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.974330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.974398 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.974412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.974433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:39:59 crc kubenswrapper[4733]: I1204 17:39:59.974451 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:39:59Z","lastTransitionTime":"2025-12-04T17:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.079673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.079714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.079726 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.079742 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.079754 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.182853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.182915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.182937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.182966 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.182988 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.285614 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.285691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.285716 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.285746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.285771 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.334663 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:00 crc kubenswrapper[4733]: E1204 17:40:00.334904 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.388940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.389001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.389018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.389041 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.389057 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.554677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.555218 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.555467 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.555693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.555940 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.659825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.660740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.660937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.661089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.661223 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.763606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.763642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.763653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.763669 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.763680 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.866349 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.866648 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.866900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.867120 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.867305 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.971038 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.971111 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.971134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.971162 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:00 crc kubenswrapper[4733]: I1204 17:40:00.971183 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:00Z","lastTransitionTime":"2025-12-04T17:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.073863 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.073900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.073912 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.073927 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.073937 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.104765 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.104817 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.104830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.104844 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.104855 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.123959 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:01Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.127459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.127488 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.127502 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.127516 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.127528 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.141049 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:01Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.145093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.145120 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.145128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.145139 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.145147 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.161211 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:01Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.165736 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.165828 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.165853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.165883 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.165904 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.186594 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:01Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.190858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.190978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.191052 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.191100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.191181 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.211309 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:01Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.211565 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.213724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.213783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.213840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.213868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.213889 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.317113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.317168 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.317184 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.317211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.317228 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.334477 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.334503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.334550 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.334676 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.335033 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.335681 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.336193 4733 scope.go:117] "RemoveContainer" containerID="4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4" Dec 04 17:40:01 crc kubenswrapper[4733]: E1204 17:40:01.336581 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.420992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.421047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.421063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.421085 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.421102 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.523428 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.523457 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.523465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.523479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.523488 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.625515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.625567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.625579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.625595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.625607 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.727853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.727897 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.727906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.727920 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.727930 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.830712 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.830755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.830768 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.830786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.830821 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.933175 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.933208 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.933220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.933234 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:01 crc kubenswrapper[4733]: I1204 17:40:01.933248 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:01Z","lastTransitionTime":"2025-12-04T17:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.035872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.035905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.035915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.035930 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.035940 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.137892 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.137928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.137939 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.137953 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.137965 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.240484 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.240513 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.240524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.240538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.240559 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.334890 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:02 crc kubenswrapper[4733]: E1204 17:40:02.335122 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.342860 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.342907 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.342918 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.342931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.342942 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.445290 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.445326 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.445338 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.445379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.445419 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.548861 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.548915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.548931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.548954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.548971 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.651344 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.651409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.651426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.651451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.651469 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.754426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.754481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.754501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.754524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.754542 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.857383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.857417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.857427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.857441 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.857455 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.960631 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.960675 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.960686 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.960702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:02 crc kubenswrapper[4733]: I1204 17:40:02.960715 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:02Z","lastTransitionTime":"2025-12-04T17:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.063424 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.063460 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.063468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.063482 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.063493 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.165741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.165829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.165840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.165852 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.165861 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.267897 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.267934 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.267945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.267960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.267972 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.334501 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.334513 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.334517 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:03 crc kubenswrapper[4733]: E1204 17:40:03.335335 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:03 crc kubenswrapper[4733]: E1204 17:40:03.335467 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:03 crc kubenswrapper[4733]: E1204 17:40:03.335607 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.370619 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.370723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.370749 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.370774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.370831 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.473196 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.473245 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.473255 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.473274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.473285 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.575434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.575479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.575492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.575521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.575535 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.678839 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.678895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.678908 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.678923 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.678936 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.781325 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.781394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.781412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.781437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.781455 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.884857 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.884904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.884917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.884932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.884944 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.989728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.989769 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.989779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.989812 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:03 crc kubenswrapper[4733]: I1204 17:40:03.989823 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:03Z","lastTransitionTime":"2025-12-04T17:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.092752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.092830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.092843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.092861 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.092872 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.198407 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.198461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.198475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.198504 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.198516 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.300758 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.300808 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.300820 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.300836 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.300846 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.335207 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:04 crc kubenswrapper[4733]: E1204 17:40:04.335366 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.403241 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.403468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.403479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.403495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.403506 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.506356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.506405 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.506417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.506461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.506473 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.608963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.609054 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.609079 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.609102 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.609122 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.711020 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.711048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.711056 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.711069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.711078 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.813878 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.813934 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.813945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.813962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.813978 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.916859 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.916897 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.916907 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.916922 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:04 crc kubenswrapper[4733]: I1204 17:40:04.916931 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:04Z","lastTransitionTime":"2025-12-04T17:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.020297 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.020355 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.020374 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.020399 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.020416 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.123056 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.123155 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.123189 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.123205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.123217 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.226608 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.226665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.226678 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.226693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.226705 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.329195 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.329250 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.329262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.329284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.329299 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.334403 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:05 crc kubenswrapper[4733]: E1204 17:40:05.334542 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.334736 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:05 crc kubenswrapper[4733]: E1204 17:40:05.334834 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.334967 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:05 crc kubenswrapper[4733]: E1204 17:40:05.335111 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.431825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.431909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.431926 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.431950 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.431967 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.534615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.534688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.534701 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.534725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.534740 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.637954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.638012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.638024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.638048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.638062 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.740142 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.740171 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.740179 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.740194 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.740203 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.843503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.843564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.843586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.843610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.843629 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.947577 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.947627 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.947645 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.947666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:05 crc kubenswrapper[4733]: I1204 17:40:05.947682 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:05Z","lastTransitionTime":"2025-12-04T17:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.050436 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.050491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.050511 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.050537 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.050556 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.153398 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.153441 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.153457 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.153479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.153496 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.256433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.256505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.256525 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.256553 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.256571 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.334601 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:06 crc kubenswrapper[4733]: E1204 17:40:06.334863 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.358544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.358583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.358591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.358606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.358618 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.461397 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.461437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.461448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.461463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.461474 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.564421 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.564492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.564515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.565373 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.565432 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.667960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.667996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.668004 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.668019 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.668028 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.770078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.770152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.770176 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.770206 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.770227 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.873367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.873455 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.873468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.873493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.873507 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.976205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.976262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.976281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.976303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:06 crc kubenswrapper[4733]: I1204 17:40:06.976321 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:06Z","lastTransitionTime":"2025-12-04T17:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.078962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.079024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.079042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.079067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.079084 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.181744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.181825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.181845 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.181868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.181886 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.284694 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.284768 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.284788 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.284850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.284875 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.335078 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.335138 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.335236 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:07 crc kubenswrapper[4733]: E1204 17:40:07.335282 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:07 crc kubenswrapper[4733]: E1204 17:40:07.335497 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:07 crc kubenswrapper[4733]: E1204 17:40:07.335640 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.387514 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.387560 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.387569 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.387586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.387596 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.490299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.490377 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.490397 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.490426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.490449 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.593093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.593140 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.593157 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.593179 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.593195 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.695606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.695642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.695653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.695666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.695676 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.798459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.798522 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.798544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.798574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.798590 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.901348 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.901383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.901394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.901407 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:07 crc kubenswrapper[4733]: I1204 17:40:07.901419 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:07Z","lastTransitionTime":"2025-12-04T17:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.003338 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.003405 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.003430 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.003459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.003484 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.105930 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.105986 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.106004 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.106028 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.106044 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.208704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.208742 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.208752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.208772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.208782 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.311293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.311384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.311401 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.311426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.311444 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.335257 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:08 crc kubenswrapper[4733]: E1204 17:40:08.335482 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.350199 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.363819 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45d07789-93de-4a10-85ee-5af716a86700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.381487 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.393842 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.406362 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.413383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.413465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.413490 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.413520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.413543 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.431128 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.441786 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.452127 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.510980 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.515599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.515635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.515646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.515662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.515673 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.529325 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.545833 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.557673 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.569998 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.579979 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.592578 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.604131 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.615122 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.617418 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.617440 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.617448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.617460 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.617469 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.628078 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:08Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.719145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.719182 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.719194 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.719209 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.719220 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.821591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.821619 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.821639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.821651 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.821659 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.923231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.923267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.923282 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.923296 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:08 crc kubenswrapper[4733]: I1204 17:40:08.923308 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:08Z","lastTransitionTime":"2025-12-04T17:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.025827 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.025864 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.025872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.025886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.025894 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.128970 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.129040 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.129058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.129082 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.129100 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.231183 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.231227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.231238 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.231256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.231270 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.334995 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.335048 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.335112 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:09 crc kubenswrapper[4733]: E1204 17:40:09.335157 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:09 crc kubenswrapper[4733]: E1204 17:40:09.335334 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:09 crc kubenswrapper[4733]: E1204 17:40:09.335444 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.362655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.362687 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.362698 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.362717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.362730 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.466303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.466344 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.466353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.466367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.466376 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.569311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.569348 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.569357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.569373 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.569383 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.671706 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.671766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.671779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.671829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.671842 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.773521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.773555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.773563 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.773575 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.773593 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.875984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.876009 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.876017 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.876029 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.876038 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.978483 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.978521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.978530 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.978544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:09 crc kubenswrapper[4733]: I1204 17:40:09.978553 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:09Z","lastTransitionTime":"2025-12-04T17:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.081203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.081273 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.081290 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.081319 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.081340 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.184110 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.184178 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.184196 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.184222 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.184240 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.234661 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:10 crc kubenswrapper[4733]: E1204 17:40:10.234880 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:40:10 crc kubenswrapper[4733]: E1204 17:40:10.234967 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs podName:79d2d07f-2f47-4970-b6e5-7d25b02e639e nodeName:}" failed. No retries permitted until 2025-12-04 17:40:42.234942164 +0000 UTC m=+104.190303250 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs") pod "network-metrics-daemon-mlwj8" (UID: "79d2d07f-2f47-4970-b6e5-7d25b02e639e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.286940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.287007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.287091 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.287132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.288050 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.334936 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:10 crc kubenswrapper[4733]: E1204 17:40:10.335187 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.391879 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.391935 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.391952 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.391978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.391996 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.495085 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.495470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.495487 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.495509 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.495526 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.598839 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.598908 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.598931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.598960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.598989 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.701283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.701333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.701344 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.701361 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.701373 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.804421 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.804489 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.804506 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.804531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.804550 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.849910 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/0.log" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.849991 4733 generic.go:334] "Generic (PLEG): container finished" podID="829fda7a-6a96-4007-b2f5-0be03782a764" containerID="8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6" exitCode=1 Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.850035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-472m8" event={"ID":"829fda7a-6a96-4007-b2f5-0be03782a764","Type":"ContainerDied","Data":"8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.850570 4733 scope.go:117] "RemoveContainer" containerID="8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.869665 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.889061 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.901171 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.914260 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.914327 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.914347 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.914376 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.914394 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:10Z","lastTransitionTime":"2025-12-04T17:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.917665 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.926724 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.941421 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:10Z\\\",\\\"message\\\":\\\"2025-12-04T17:39:25+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b\\\\n2025-12-04T17:39:25+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b to /host/opt/cni/bin/\\\\n2025-12-04T17:39:25Z [verbose] multus-daemon started\\\\n2025-12-04T17:39:25Z [verbose] Readiness Indicator file check\\\\n2025-12-04T17:40:10Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.951883 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.965391 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:10 crc kubenswrapper[4733]: I1204 17:40:10.988116 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:10Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.006766 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.019364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.019412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.019429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.019453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.019470 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.021690 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45d07789-93de-4a10-85ee-5af716a86700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.056679 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.073428 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.093226 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.116168 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.121145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.121165 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.121174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.121187 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.121196 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.128045 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.141431 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.164325 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.223188 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.223221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.223231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.223246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.223257 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.266829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.266890 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.266908 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.266930 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.266947 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.281642 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.286265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.286286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.286294 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.286308 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.286317 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.305010 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.309001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.309025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.309050 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.309061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.309069 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.334405 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.334476 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.334507 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.334634 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.334655 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.334697 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.338355 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.343569 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.343623 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.343646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.343673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.343701 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.362979 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.366779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.366862 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.366880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.366904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.366924 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.384927 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: E1204 17:40:11.385164 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.386917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.386961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.386976 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.386997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.387011 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.489135 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.489185 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.489201 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.489223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.489240 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.591354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.591407 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.591425 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.591448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.591465 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.693543 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.693638 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.693657 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.693682 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.693741 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.796678 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.796718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.796755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.796770 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.796779 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.854938 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/0.log" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.855016 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-472m8" event={"ID":"829fda7a-6a96-4007-b2f5-0be03782a764","Type":"ContainerStarted","Data":"dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.873552 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45d07789-93de-4a10-85ee-5af716a86700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.893108 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.899669 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.899714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.899730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.899752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.899770 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:11Z","lastTransitionTime":"2025-12-04T17:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.915282 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.929881 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.946892 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:11 crc kubenswrapper[4733]: I1204 17:40:11.979503 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.000216 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:11Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.003093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.003159 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.003175 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.003195 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.003210 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.020324 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.039126 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.069747 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.088480 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.106134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.106189 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.106216 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.106245 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.106267 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.109468 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.131181 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.149826 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.172256 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.186298 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.202132 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.219323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.219415 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.219438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.219462 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.219519 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.224172 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:10Z\\\",\\\"message\\\":\\\"2025-12-04T17:39:25+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b\\\\n2025-12-04T17:39:25+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b to /host/opt/cni/bin/\\\\n2025-12-04T17:39:25Z [verbose] multus-daemon started\\\\n2025-12-04T17:39:25Z [verbose] Readiness Indicator file check\\\\n2025-12-04T17:40:10Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:12Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.323001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.323033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.323041 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.323057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.323067 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.334638 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:12 crc kubenswrapper[4733]: E1204 17:40:12.334893 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.426462 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.435154 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.435273 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.435307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.435382 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.539261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.539312 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.539328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.539353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.539371 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.642014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.642132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.642188 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.642212 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.642231 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.745887 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.745937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.745958 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.745986 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.746007 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.849905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.849999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.850015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.850036 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.850052 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.953076 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.953126 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.953143 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.953166 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:12 crc kubenswrapper[4733]: I1204 17:40:12.953185 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:12Z","lastTransitionTime":"2025-12-04T17:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.056363 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.056428 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.056446 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.056478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.056496 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.159464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.159538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.159562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.159592 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.159614 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.262482 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.262544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.262561 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.262586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.262602 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.334907 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.334923 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:13 crc kubenswrapper[4733]: E1204 17:40:13.335126 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:13 crc kubenswrapper[4733]: E1204 17:40:13.335283 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.335745 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:13 crc kubenswrapper[4733]: E1204 17:40:13.336111 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.365722 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.365874 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.365900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.365927 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.365950 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.468517 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.468853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.469237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.469544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.469880 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.573045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.573142 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.573162 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.573262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.573350 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.678268 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.678304 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.678316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.678334 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.678346 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.781498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.781835 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.782013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.782224 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.782377 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.885329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.885688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.885871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.886017 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.886175 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.989704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.989787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.989862 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.989889 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:13 crc kubenswrapper[4733]: I1204 17:40:13.989905 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:13Z","lastTransitionTime":"2025-12-04T17:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.093067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.093133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.093151 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.093176 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.093193 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.196026 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.196081 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.196096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.196121 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.196138 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.298846 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.298910 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.298932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.298959 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.298981 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.335132 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:14 crc kubenswrapper[4733]: E1204 17:40:14.335318 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.402293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.402331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.402339 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.402356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.402364 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.505503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.505572 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.505591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.505617 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.505636 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.608737 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.608780 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.608829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.608854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.608866 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.712560 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.712616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.712633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.712656 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.712675 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.814957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.815001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.815012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.815029 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.815039 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.917870 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.917915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.917925 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.917941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:14 crc kubenswrapper[4733]: I1204 17:40:14.917953 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:14Z","lastTransitionTime":"2025-12-04T17:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.021453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.021496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.021509 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.021525 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.021537 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.125881 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.125945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.125962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.125987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.126005 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.229873 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.229995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.230058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.230095 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.230155 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.333762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.333862 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.333880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.333903 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.333923 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.334682 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.334701 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.334854 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:15 crc kubenswrapper[4733]: E1204 17:40:15.334909 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:15 crc kubenswrapper[4733]: E1204 17:40:15.335082 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:15 crc kubenswrapper[4733]: E1204 17:40:15.335175 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.437521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.437583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.437600 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.437626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.437644 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.542685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.542973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.543037 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.543099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.543180 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.646536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.646607 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.646626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.646653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.646673 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.749689 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.749753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.749774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.749832 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.749852 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.852880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.852945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.852964 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.852987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.853005 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.956309 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.956367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.956384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.956411 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:15 crc kubenswrapper[4733]: I1204 17:40:15.956430 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:15Z","lastTransitionTime":"2025-12-04T17:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.061646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.061701 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.061718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.061740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.061760 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.164261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.164353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.164405 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.164428 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.164445 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.266962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.267007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.267020 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.267043 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.267057 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.336152 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:16 crc kubenswrapper[4733]: E1204 17:40:16.336780 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.337152 4733 scope.go:117] "RemoveContainer" containerID="4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.369461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.369490 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.369498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.369511 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.369520 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.471980 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.472027 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.472044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.472069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.472088 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.575667 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.575711 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.575723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.575740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.575755 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.678904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.678983 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.679009 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.679042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.679071 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.781400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.781451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.781465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.781486 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.781502 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.875079 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/2.log" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.879350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.879962 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.883491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.883520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.883529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.883540 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.883549 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.894250 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:10Z\\\",\\\"message\\\":\\\"2025-12-04T17:39:25+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b\\\\n2025-12-04T17:39:25+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b to /host/opt/cni/bin/\\\\n2025-12-04T17:39:25Z [verbose] multus-daemon started\\\\n2025-12-04T17:39:25Z [verbose] Readiness Indicator file check\\\\n2025-12-04T17:40:10Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:16Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.915558 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:16Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.930660 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:16Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.948423 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:16Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.965668 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:16Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.985840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.985889 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.985900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.985919 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.985931 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:16Z","lastTransitionTime":"2025-12-04T17:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:16 crc kubenswrapper[4733]: I1204 17:40:16.989616 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45d07789-93de-4a10-85ee-5af716a86700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:16Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.011582 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.037247 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.050703 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.067104 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:40:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.076132 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.088024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.088149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.088076 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.088242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.088567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.088650 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.107970 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.123147 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.136889 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.149353 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.164233 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.176250 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.190765 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.190809 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.190818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.190832 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.190842 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.293517 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.293569 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.293586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.293606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.293624 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.335411 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.335458 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.335938 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:17 crc kubenswrapper[4733]: E1204 17:40:17.336223 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:17 crc kubenswrapper[4733]: E1204 17:40:17.336113 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:17 crc kubenswrapper[4733]: E1204 17:40:17.336535 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.396110 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.396329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.396391 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.396459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.396522 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.500262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.500301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.500315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.500333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.500347 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.603127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.603166 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.603174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.603186 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.603196 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.704630 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.704937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.705078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.705200 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.705334 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.808138 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.808187 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.808204 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.808224 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.808241 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.886404 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/3.log" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.887430 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/2.log" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.891900 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" exitCode=1 Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.891965 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.892017 4733 scope.go:117] "RemoveContainer" containerID="4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.893331 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:40:17 crc kubenswrapper[4733]: E1204 17:40:17.894013 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.911014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.911081 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.911106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.911137 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.911159 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:17Z","lastTransitionTime":"2025-12-04T17:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.915360 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.930954 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.951312 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.973223 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:17 crc kubenswrapper[4733]: I1204 17:40:17.991956 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:17Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.011056 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.016423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.016487 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.016510 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.016541 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.016567 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.033190 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:10Z\\\",\\\"message\\\":\\\"2025-12-04T17:39:25+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b\\\\n2025-12-04T17:39:25+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b to /host/opt/cni/bin/\\\\n2025-12-04T17:39:25Z [verbose] multus-daemon started\\\\n2025-12-04T17:39:25Z [verbose] Readiness Indicator file check\\\\n2025-12-04T17:40:10Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.054058 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45d07789-93de-4a10-85ee-5af716a86700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.070347 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.093078 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.110203 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.119219 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.119252 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.119261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.119276 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.119284 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.139477 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:17Z\\\",\\\"message\\\":\\\"rator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:40:17.404302 6728 factory.go:656] Stopping watch factory\\\\nI1204 17:40:17.404331 6728 ovnkube.go:599] Stopped ovnkube\\\\nI1204 17:40:17.404395 6728 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 17:40:17.404413 6728 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nI1204 17:40:17.404441 6728 services_controller.go:360] Finished syncing service check-endpoints on namespace openshift-apiserver for network=default : 1.12285ms\\\\nI1204 17:40:17.404467 6728 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nF1204 17:40:17.404477 6728 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:40:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.156304 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.168429 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.188342 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.200968 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.213483 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.225768 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.225829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.225841 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.225857 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.225869 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.229109 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.329203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.329260 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.329277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.329300 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.329317 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.334977 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:18 crc kubenswrapper[4733]: E1204 17:40:18.335184 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.351885 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.375911 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.392611 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.410123 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45d07789-93de-4a10-85ee-5af716a86700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.431987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.432042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.432055 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.432077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.432090 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.432614 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.451942 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.468578 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.498185 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a9e8f15d72c830bc2177eb65ac06d8bf4a08e584030bbea54cdc5c7aeafcfc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:39:49Z\\\",\\\"message\\\":\\\"sing sync for service openshift-console/console for network=default\\\\nI1204 17:39:49.268468 6370 services_controller.go:451] Built service default/kubernetes cluster-wide LB for network=default: []services.LB{}\\\\nI1204 17:39:49.268603 6370 services_controller.go:452] Built service default/kubernetes per-node LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:17Z\\\",\\\"message\\\":\\\"rator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:40:17.404302 6728 factory.go:656] Stopping watch factory\\\\nI1204 17:40:17.404331 6728 ovnkube.go:599] Stopped ovnkube\\\\nI1204 17:40:17.404395 6728 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 17:40:17.404413 6728 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nI1204 17:40:17.404441 6728 services_controller.go:360] Finished syncing service check-endpoints on namespace openshift-apiserver for network=default : 1.12285ms\\\\nI1204 17:40:17.404467 6728 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nF1204 17:40:17.404477 6728 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:40:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.515004 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.532488 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.535178 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.535290 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.535313 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.535372 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.535450 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.565526 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.585921 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.607511 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.625726 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.638368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.638444 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.638465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.638495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.638518 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.646168 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.660568 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.678559 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:10Z\\\",\\\"message\\\":\\\"2025-12-04T17:39:25+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b\\\\n2025-12-04T17:39:25+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b to /host/opt/cni/bin/\\\\n2025-12-04T17:39:25Z [verbose] multus-daemon started\\\\n2025-12-04T17:39:25Z [verbose] Readiness Indicator file check\\\\n2025-12-04T17:40:10Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.695422 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.741905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.742007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.742027 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.742058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.742077 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.845286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.845417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.845443 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.845475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.845498 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.898640 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/3.log" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.910146 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:40:18 crc kubenswrapper[4733]: E1204 17:40:18.910400 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.928374 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70e27a7c6079c747ae79dcdec2711a1bb7b9388362ff060e752112391f26be50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9pcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w7n9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.949000 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.949064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.949086 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.949119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.949141 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:18Z","lastTransitionTime":"2025-12-04T17:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.950081 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-472m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"829fda7a-6a96-4007-b2f5-0be03782a764\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:10Z\\\",\\\"message\\\":\\\"2025-12-04T17:39:25+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b\\\\n2025-12-04T17:39:25+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1c46a72d-b331-44b3-a750-70836389d40b to /host/opt/cni/bin/\\\\n2025-12-04T17:39:25Z [verbose] multus-daemon started\\\\n2025-12-04T17:39:25Z [verbose] Readiness Indicator file check\\\\n2025-12-04T17:40:10Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swzhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-472m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.969899 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45d07789-93de-4a10-85ee-5af716a86700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8ce0a8680303600adab1a3aca20ea77dd3ce413c2c38df301d765d8c03fa8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e84506681b9127307759fbd64ba0def1a24c5a8782c8a9b4011aa27bd16cc28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8af4ba5eb62bb92b461c6798049605b61e4179d3ecd0416333c6d1f4a06b0ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7173984e0c2c281a06a1ad82febd29170dda848d06a0c824f1adab02a4fb660\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:18 crc kubenswrapper[4733]: I1204 17:40:18.987850 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:18Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.007896 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ef97472-7ccc-4d3d-a127-e70bced45051\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb992a9c536141304adbd806d006043d7a2e13fd75f25259726dc247a6f6344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435c1f219fe708b980ade555501251792f1695f5d4c624aef2d836907109c4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cae4d2bd431f0067ddd1517896acdb3e456d89fb80d7bf0737ed332fbbc1fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2464d47f387e1024d2df4a29e6922a2f4ba64d165c82dd0a6b75256ccc27f442\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91e8d40db674bbdbc31485265a94ebe59a84b011da3002030e109155de0c0f24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98dcfe0432a2152005503a3ef3f867426f73f61ca7d8c3901e4996acb2895066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7027166c4ba6523fdeac9e2b2503c95120326e01efb26dc4613cf196646cc8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj56j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xf8lf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.022116 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d2d07f-2f47-4970-b6e5-7d25b02e639e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kp2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mlwj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.044690 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"885d8e44-847b-4e14-b6b5-df3fb0d445d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7d57b12ab70a6386087ca2b4db27a4bc152bc4a2cdfbf87876be2d4264fa407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32779a85679df23b6d547a191c0621aba6153e90830f7771479beee228025243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3278436d58fe0369570b60804e312e25f4c1c5d9d976e93411bce1525e6c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd3b9833d6c5c57267f287726289ef47102a8ed2bf673276727f010e3af6550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://944088b187ea3ece437bc190a3bb6ea4cff9e9b47d54f020b4fa02c4f8f3bfca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3cb41ed19c457b59e93535bb28563ae3264ba7c8414dea61cef1b872728590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e8f10b2b729db88f4e716801c2ae27fa8ba2986f7174434ea6388942e737c9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47568644d7677be8f41700c7700a19c1744d221cb664840740c34441e159598d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.051232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.051354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.051375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.051399 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.051417 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.063108 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cceba9-c503-4d25-8c75-b634e1bb59d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T17:39:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1204 17:39:10.969203 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 17:39:10.970864 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4014136482/tls.crt::/tmp/serving-cert-4014136482/tls.key\\\\\\\"\\\\nI1204 17:39:16.337503 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 17:39:16.340866 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 17:39:16.340926 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 17:39:16.340969 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 17:39:16.340985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 17:39:16.351062 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 17:39:16.351101 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 17:39:16.351125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 17:39:16.351136 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 17:39:16.351144 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 17:39:16.351154 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 17:39:16.351501 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 17:39:16.353848 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:38:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.083557 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.102542 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.134913 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"067d57d5-c6ec-463d-b1a0-ec5e8c686453\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T17:40:17Z\\\",\\\"message\\\":\\\"rator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 17:40:17.404302 6728 factory.go:656] Stopping watch factory\\\\nI1204 17:40:17.404331 6728 ovnkube.go:599] Stopped ovnkube\\\\nI1204 17:40:17.404395 6728 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 17:40:17.404413 6728 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}\\\\nI1204 17:40:17.404441 6728 services_controller.go:360] Finished syncing service check-endpoints on namespace openshift-apiserver for network=default : 1.12285ms\\\\nI1204 17:40:17.404467 6728 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nF1204 17:40:17.404477 6728 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T17:40:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T17:39:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vhdsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qp9sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.153564 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sl9kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660f910c-5b41-457f-aeda-3b6319b67533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6304d1b05dd15132da74b9bd4c16b231c9bd3a95e626cfb8e7a3d45bb50e54b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx5hd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sl9kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.154953 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.155008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.155026 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.155052 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.155070 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.172023 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00bca008-075a-4027-944f-b3c9cad33814\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0afcf50c45bd9e1e5a040ce741c26902043611e77ad0e3c68a44eb1c5ea8a433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05004817cba8853902aa317ab886614868dece82158b4feb3c1b3cb87b7e3b13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nwhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rv2n2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.191640 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f00d00db-963d-4197-8f82-501a5926f21c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:38:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f26bbabd325a74724a057cc0273d0a6a3cf4a57b23af501fbbbf48347b0b49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://580783787fe5f4694311f9ce8c0a718334dabfe89b9a2e52530e597375f96913\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:38:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccdf638216d8129ada7b7f6b4b89fac81b5fdb16bb3d1a0fb6b52ebcc0488b8c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:38:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.218548 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://553a948d272bce0e49e267965a3ca1b3436666aa67bed6207642b11db9b62334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.238950 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bdb7cc9bc6a68217f10af2e59e86d12fc09e599e5f2456220166184869bc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.259582 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4151d9e68db53aaa9bcb500c206070a111454e1d949c05ecd20f2077c0f9702e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c18f843a05afde9e6924265038928a4fa306f10a0592ec73840fc2b3bc3df326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.259678 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.259723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.259743 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.259771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.259789 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.272570 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6tl82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21f90af6-f871-4cce-88b4-824c4a15e42c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T17:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://322c747987daa11f82e70a4e1fa9ba8ac202ea191520113f1db87f5eec4c25cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T17:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwgwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T17:39:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6tl82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:19Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.334452 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.334574 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.334451 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:19 crc kubenswrapper[4733]: E1204 17:40:19.334649 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:19 crc kubenswrapper[4733]: E1204 17:40:19.334770 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:19 crc kubenswrapper[4733]: E1204 17:40:19.334973 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.362869 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.362950 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.362974 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.363006 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.363034 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.466107 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.466201 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.466220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.466244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.466262 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.569341 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.569464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.569484 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.569507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.569523 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.673122 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.673166 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.673183 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.673205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.673221 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.776457 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.776521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.776545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.776573 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.776590 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.879325 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.879381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.879401 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.879426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.879444 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.981960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.982010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.982026 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.982047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:19 crc kubenswrapper[4733]: I1204 17:40:19.982064 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:19Z","lastTransitionTime":"2025-12-04T17:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.085331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.085384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.085400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.085422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.085439 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.188501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.188551 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.188568 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.188591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.188608 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.291884 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.291933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.291952 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.291979 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.292065 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.395403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.395446 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.395462 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.395487 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.395509 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.462590 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:20 crc kubenswrapper[4733]: E1204 17:40:20.462766 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.468317 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:20 crc kubenswrapper[4733]: E1204 17:40:20.468417 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.504204 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.504256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.504271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.504292 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.504309 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.607258 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.607315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.607332 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.607357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.607374 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.710224 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.710306 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.710333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.710364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.710386 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.813995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.814059 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.814076 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.814102 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.814126 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.916415 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.916480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.916498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.916522 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:20 crc kubenswrapper[4733]: I1204 17:40:20.916542 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:20Z","lastTransitionTime":"2025-12-04T17:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.019888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.019944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.019961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.019985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.020002 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.123462 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.123520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.123536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.123560 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.123578 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.227199 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.227282 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.227299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.227322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.227338 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.250928 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.251280 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.251206847 +0000 UTC m=+147.206567933 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.330961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.331051 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.331070 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.331094 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.331113 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.335384 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.335473 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.335556 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.335652 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.352093 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.352156 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.352207 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.352269 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352277 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352287 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352315 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352345 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352392 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.352364508 +0000 UTC m=+147.307725594 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352418 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352422 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.352408569 +0000 UTC m=+147.307769645 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352520 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.352493961 +0000 UTC m=+147.307855037 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352422 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352552 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352571 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.352625 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.352612724 +0000 UTC m=+147.307973810 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.433621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.433710 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.433728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.433782 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.433820 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.457418 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.457475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.457492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.457513 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.457532 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.478778 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.484068 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.484136 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.484153 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.484179 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.484197 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.504542 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.509546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.509593 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.509610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.509635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.509653 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.529219 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.534252 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.534321 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.534338 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.534364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.534381 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.555403 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.560738 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.560836 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.560857 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.560882 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.560901 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.580834 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T17:40:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d01826a6-fc94-4cf3-8ea3-4b2d5fb4094f\\\",\\\"systemUUID\\\":\\\"61d6cd1d-2e94-4088-ae6a-14e04c4bcca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T17:40:21Z is after 2025-08-24T17:21:41Z" Dec 04 17:40:21 crc kubenswrapper[4733]: E1204 17:40:21.581058 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.583495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.583540 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.583558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.583579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.583595 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.685729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.685783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.685841 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.685862 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.685881 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.788681 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.788775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.788854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.788880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.788897 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.892277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.892335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.892361 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.892393 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.892416 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.995854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.995914 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.995941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.995969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:21 crc kubenswrapper[4733]: I1204 17:40:21.995990 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:21Z","lastTransitionTime":"2025-12-04T17:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.099117 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.099233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.099257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.099286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.099309 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.202616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.202679 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.202695 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.202723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.202745 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.310680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.310790 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.310870 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.310904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.310921 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.334547 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:22 crc kubenswrapper[4733]: E1204 17:40:22.334722 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.334770 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:22 crc kubenswrapper[4733]: E1204 17:40:22.334969 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.413610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.413663 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.413680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.413701 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.413718 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.516980 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.517068 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.517092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.517126 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.517156 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.620312 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.620371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.620389 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.620414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.620434 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.723102 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.723158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.723174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.723195 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.723211 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.827003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.827566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.827953 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.828281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.828545 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.932058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.932116 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.932160 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.932185 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:22 crc kubenswrapper[4733]: I1204 17:40:22.932202 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:22Z","lastTransitionTime":"2025-12-04T17:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.035487 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.036163 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.036430 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.036597 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.036845 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.140571 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.140860 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.141105 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.141289 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.141455 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.244479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.244537 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.244553 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.244580 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.244599 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.334530 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.334610 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:23 crc kubenswrapper[4733]: E1204 17:40:23.334763 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:23 crc kubenswrapper[4733]: E1204 17:40:23.335053 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.348178 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.348436 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.348637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.348894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.349081 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.452082 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.452611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.452998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.453368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.453723 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.556779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.556904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.556927 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.556952 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.556971 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.659979 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.660045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.660065 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.660096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.660116 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.764332 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.764429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.764454 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.764484 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.764505 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.868024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.868090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.868147 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.868179 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.868199 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.970900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.970949 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.970965 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.970989 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:23 crc kubenswrapper[4733]: I1204 17:40:23.971007 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:23Z","lastTransitionTime":"2025-12-04T17:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.079317 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.079376 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.079393 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.079414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.079432 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.182335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.182427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.182452 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.182485 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.182508 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.285691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.285767 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.285843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.285996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.286053 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.334621 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:24 crc kubenswrapper[4733]: E1204 17:40:24.334778 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.334898 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:24 crc kubenswrapper[4733]: E1204 17:40:24.335324 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.352341 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.388888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.388997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.389070 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.389107 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.389184 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.492280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.492333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.492356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.492378 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.492394 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.596256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.596314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.596331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.596356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.596373 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.699275 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.699358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.699383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.699414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.699438 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.802978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.803080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.803098 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.803155 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.803173 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.905884 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.905957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.905978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.906002 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:24 crc kubenswrapper[4733]: I1204 17:40:24.906019 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:24Z","lastTransitionTime":"2025-12-04T17:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.010096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.010160 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.010183 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.010209 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.010228 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.114100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.114171 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.114195 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.114227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.114250 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.218164 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.218240 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.218261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.218284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.218301 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.320995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.321055 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.321078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.321106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.321128 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.334428 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.334514 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:25 crc kubenswrapper[4733]: E1204 17:40:25.334591 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:25 crc kubenswrapper[4733]: E1204 17:40:25.334663 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.423566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.423632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.423654 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.423681 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.423702 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.527300 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.527379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.527406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.527438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.527458 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.630054 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.630126 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.630145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.630172 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.630192 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.733545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.733653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.733671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.733697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.733714 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.837236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.837300 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.837316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.837343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.837360 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.940469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.940601 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.940681 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.940713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:25 crc kubenswrapper[4733]: I1204 17:40:25.940731 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:25Z","lastTransitionTime":"2025-12-04T17:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.043997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.044063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.044084 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.044111 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.044131 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.146290 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.146395 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.146414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.146437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.146454 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.248986 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.249055 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.249074 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.249099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.249118 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.335142 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.335280 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:26 crc kubenswrapper[4733]: E1204 17:40:26.335336 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:26 crc kubenswrapper[4733]: E1204 17:40:26.335501 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.352033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.352181 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.352210 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.352241 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.352265 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.455272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.455353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.455370 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.455401 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.455418 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.559097 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.559154 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.559174 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.559197 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.559213 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.661349 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.661387 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.661395 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.661408 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.661418 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.763689 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.763717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.763725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.763737 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.763746 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.867128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.867165 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.867173 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.867208 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.867220 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.970321 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.970375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.970384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.970401 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:26 crc kubenswrapper[4733]: I1204 17:40:26.970416 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:26Z","lastTransitionTime":"2025-12-04T17:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.078818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.078886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.078918 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.078947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.078966 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.182070 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.182501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.182519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.182542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.182560 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.285459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.285561 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.285582 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.285605 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.285660 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.334641 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.334655 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:27 crc kubenswrapper[4733]: E1204 17:40:27.334839 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:27 crc kubenswrapper[4733]: E1204 17:40:27.334987 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.388062 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.388124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.388143 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.388170 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.388187 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.490943 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.490997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.491014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.491037 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.491054 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.594268 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.594422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.594449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.594481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.594507 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.696996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.697089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.697108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.697133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.697153 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.802097 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.802217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.802237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.802267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.802283 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.905327 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.905396 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.905418 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.905449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:27 crc kubenswrapper[4733]: I1204 17:40:27.905472 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:27Z","lastTransitionTime":"2025-12-04T17:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.008635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.008696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.008714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.008737 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.008754 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.111662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.111724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.111742 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.111766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.111787 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.214123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.214186 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.214211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.214240 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.214262 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.317291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.317358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.317375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.317400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.317423 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.334872 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.334968 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:28 crc kubenswrapper[4733]: E1204 17:40:28.335107 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:28 crc kubenswrapper[4733]: E1204 17:40:28.335269 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.385186 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=68.385127196 podStartE2EDuration="1m8.385127196s" podCreationTimestamp="2025-12-04 17:39:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.384547151 +0000 UTC m=+90.339908267" watchObservedRunningTime="2025-12-04 17:40:28.385127196 +0000 UTC m=+90.340488282" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.412056 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=72.41202935 podStartE2EDuration="1m12.41202935s" podCreationTimestamp="2025-12-04 17:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.411578967 +0000 UTC m=+90.366940093" watchObservedRunningTime="2025-12-04 17:40:28.41202935 +0000 UTC m=+90.367390436" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.420898 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.421239 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.421447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.421590 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.421723 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.522697 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-sl9kt" podStartSLOduration=65.522676494 podStartE2EDuration="1m5.522676494s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.52215845 +0000 UTC m=+90.477519586" watchObservedRunningTime="2025-12-04 17:40:28.522676494 +0000 UTC m=+90.478037560" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.523901 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.523950 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.523966 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.524337 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.524372 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.537649 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rv2n2" podStartSLOduration=64.537632356 podStartE2EDuration="1m4.537632356s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.536331131 +0000 UTC m=+90.491692197" watchObservedRunningTime="2025-12-04 17:40:28.537632356 +0000 UTC m=+90.492993412" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.583026 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=4.582993646 podStartE2EDuration="4.582993646s" podCreationTimestamp="2025-12-04 17:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.558178109 +0000 UTC m=+90.513539175" watchObservedRunningTime="2025-12-04 17:40:28.582993646 +0000 UTC m=+90.538354732" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.606579 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=70.60655794 podStartE2EDuration="1m10.60655794s" podCreationTimestamp="2025-12-04 17:39:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.584333512 +0000 UTC m=+90.539694578" watchObservedRunningTime="2025-12-04 17:40:28.60655794 +0000 UTC m=+90.561918996" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.627988 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.628028 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.628039 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.628058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.628070 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.653545 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-6tl82" podStartSLOduration=65.653524303 podStartE2EDuration="1m5.653524303s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.651595101 +0000 UTC m=+90.606956157" watchObservedRunningTime="2025-12-04 17:40:28.653524303 +0000 UTC m=+90.608885359" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.692918 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podStartSLOduration=65.692890591 podStartE2EDuration="1m5.692890591s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.670398596 +0000 UTC m=+90.625759682" watchObservedRunningTime="2025-12-04 17:40:28.692890591 +0000 UTC m=+90.648251667" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.693730 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-472m8" podStartSLOduration=65.693721223 podStartE2EDuration="1m5.693721223s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.692713316 +0000 UTC m=+90.648074402" watchObservedRunningTime="2025-12-04 17:40:28.693721223 +0000 UTC m=+90.649082299" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.718197 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=35.718172011 podStartE2EDuration="35.718172011s" podCreationTimestamp="2025-12-04 17:39:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.716192107 +0000 UTC m=+90.671553243" watchObservedRunningTime="2025-12-04 17:40:28.718172011 +0000 UTC m=+90.673533097" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.731100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.731180 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.731255 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.731278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.731296 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.800466 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xf8lf" podStartSLOduration=65.800444573 podStartE2EDuration="1m5.800444573s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:28.799545129 +0000 UTC m=+90.754906165" watchObservedRunningTime="2025-12-04 17:40:28.800444573 +0000 UTC m=+90.755805629" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.834516 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.834568 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.834584 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.834611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.834631 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.937158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.937227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.937244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.937269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:28 crc kubenswrapper[4733]: I1204 17:40:28.937287 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:28Z","lastTransitionTime":"2025-12-04T17:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.040982 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.041075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.041102 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.041131 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.041148 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.143439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.144741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.144760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.144776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.144789 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.247878 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.247937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.247959 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.247991 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.248014 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.336267 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:29 crc kubenswrapper[4733]: E1204 17:40:29.336356 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.336523 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:29 crc kubenswrapper[4733]: E1204 17:40:29.336566 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.351626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.351664 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.351676 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.351693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.351707 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.455371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.455442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.455460 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.455486 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.455505 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.564862 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.564923 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.564940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.564964 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.564982 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.668254 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.668315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.668331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.668358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.668382 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.771465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.771524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.771543 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.771566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.771583 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.875119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.875459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.875595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.875716 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.875968 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.980967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.981059 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.981077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.981099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:29 crc kubenswrapper[4733]: I1204 17:40:29.981117 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:29Z","lastTransitionTime":"2025-12-04T17:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.084691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.084832 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.084850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.084877 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.084893 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.188012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.188096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.188118 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.188144 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.188161 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.291361 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.291417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.291429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.291447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.291459 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.335027 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.335051 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:30 crc kubenswrapper[4733]: E1204 17:40:30.335256 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:30 crc kubenswrapper[4733]: E1204 17:40:30.335465 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.394479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.394539 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.394556 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.394585 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.394602 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.497465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.497528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.497545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.497568 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.497585 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.600293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.600343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.600360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.600383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.600402 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.703006 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.703063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.703080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.703121 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.703139 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.806237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.806305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.806330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.806361 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.806383 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.909422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.909485 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.909502 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.909524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:30 crc kubenswrapper[4733]: I1204 17:40:30.909540 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:30Z","lastTransitionTime":"2025-12-04T17:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.012892 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.012958 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.012976 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.013002 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.013021 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:31Z","lastTransitionTime":"2025-12-04T17:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.115265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.115303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.115320 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.115342 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.115360 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:31Z","lastTransitionTime":"2025-12-04T17:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.218264 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.218357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.218378 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.218402 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.218420 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:31Z","lastTransitionTime":"2025-12-04T17:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.320897 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.320967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.320993 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.321021 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.321044 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:31Z","lastTransitionTime":"2025-12-04T17:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.334360 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:31 crc kubenswrapper[4733]: E1204 17:40:31.334513 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.334590 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:31 crc kubenswrapper[4733]: E1204 17:40:31.334970 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.424473 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.424540 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.424557 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.424581 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.424599 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:31Z","lastTransitionTime":"2025-12-04T17:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.527351 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.527412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.527432 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.527456 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.527472 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:31Z","lastTransitionTime":"2025-12-04T17:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.612575 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.612648 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.612671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.612742 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.612768 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:31Z","lastTransitionTime":"2025-12-04T17:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.641033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.641108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.641135 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.641167 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.641192 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T17:40:31Z","lastTransitionTime":"2025-12-04T17:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.683458 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9"] Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.683983 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.689639 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.690900 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.691107 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.691763 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.784097 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.784159 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.784279 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.784333 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.784363 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.884993 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.885082 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.885113 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.885166 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.885199 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.885287 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.885343 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.899581 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.899603 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:31 crc kubenswrapper[4733]: I1204 17:40:31.914977 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60d7c50-52ed-4ab2-b262-1eadb45d9b9f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9mln9\" (UID: \"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:32 crc kubenswrapper[4733]: I1204 17:40:32.012293 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" Dec 04 17:40:32 crc kubenswrapper[4733]: W1204 17:40:32.032768 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb60d7c50_52ed_4ab2_b262_1eadb45d9b9f.slice/crio-0783444fcd8d66706714d670f8ffcf756d066cf16b5f3b448a2feae8702ac373 WatchSource:0}: Error finding container 0783444fcd8d66706714d670f8ffcf756d066cf16b5f3b448a2feae8702ac373: Status 404 returned error can't find the container with id 0783444fcd8d66706714d670f8ffcf756d066cf16b5f3b448a2feae8702ac373 Dec 04 17:40:32 crc kubenswrapper[4733]: I1204 17:40:32.335418 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:32 crc kubenswrapper[4733]: E1204 17:40:32.335580 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:32 crc kubenswrapper[4733]: I1204 17:40:32.335678 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:32 crc kubenswrapper[4733]: E1204 17:40:32.335986 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:32 crc kubenswrapper[4733]: I1204 17:40:32.336840 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:40:32 crc kubenswrapper[4733]: E1204 17:40:32.337361 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:40:32 crc kubenswrapper[4733]: I1204 17:40:32.515458 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" event={"ID":"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f","Type":"ContainerStarted","Data":"99010ab46937becc39c922be38b630c21db379d53cfcb4da603c3322251e63e6"} Dec 04 17:40:32 crc kubenswrapper[4733]: I1204 17:40:32.515752 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" event={"ID":"b60d7c50-52ed-4ab2-b262-1eadb45d9b9f","Type":"ContainerStarted","Data":"0783444fcd8d66706714d670f8ffcf756d066cf16b5f3b448a2feae8702ac373"} Dec 04 17:40:33 crc kubenswrapper[4733]: I1204 17:40:33.335004 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:33 crc kubenswrapper[4733]: E1204 17:40:33.335208 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:33 crc kubenswrapper[4733]: I1204 17:40:33.335004 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:33 crc kubenswrapper[4733]: E1204 17:40:33.335616 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:34 crc kubenswrapper[4733]: I1204 17:40:34.335385 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:34 crc kubenswrapper[4733]: I1204 17:40:34.335875 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:34 crc kubenswrapper[4733]: E1204 17:40:34.336015 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:34 crc kubenswrapper[4733]: E1204 17:40:34.336116 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:35 crc kubenswrapper[4733]: I1204 17:40:35.334367 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:35 crc kubenswrapper[4733]: I1204 17:40:35.334440 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:35 crc kubenswrapper[4733]: E1204 17:40:35.334532 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:35 crc kubenswrapper[4733]: E1204 17:40:35.334652 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:36 crc kubenswrapper[4733]: I1204 17:40:36.334707 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:36 crc kubenswrapper[4733]: I1204 17:40:36.334918 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:36 crc kubenswrapper[4733]: E1204 17:40:36.334974 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:36 crc kubenswrapper[4733]: E1204 17:40:36.335117 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:37 crc kubenswrapper[4733]: I1204 17:40:37.335083 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:37 crc kubenswrapper[4733]: E1204 17:40:37.335312 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:37 crc kubenswrapper[4733]: I1204 17:40:37.335375 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:37 crc kubenswrapper[4733]: E1204 17:40:37.335622 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:38 crc kubenswrapper[4733]: I1204 17:40:38.335044 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:38 crc kubenswrapper[4733]: I1204 17:40:38.335165 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:38 crc kubenswrapper[4733]: E1204 17:40:38.336857 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:38 crc kubenswrapper[4733]: E1204 17:40:38.337079 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:39 crc kubenswrapper[4733]: I1204 17:40:39.334905 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:39 crc kubenswrapper[4733]: I1204 17:40:39.334960 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:39 crc kubenswrapper[4733]: E1204 17:40:39.335076 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:39 crc kubenswrapper[4733]: E1204 17:40:39.335188 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:40 crc kubenswrapper[4733]: I1204 17:40:40.335171 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:40 crc kubenswrapper[4733]: I1204 17:40:40.335197 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:40 crc kubenswrapper[4733]: E1204 17:40:40.335379 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:40 crc kubenswrapper[4733]: E1204 17:40:40.335525 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:41 crc kubenswrapper[4733]: I1204 17:40:41.334856 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:41 crc kubenswrapper[4733]: I1204 17:40:41.334926 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:41 crc kubenswrapper[4733]: E1204 17:40:41.335050 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:41 crc kubenswrapper[4733]: E1204 17:40:41.335248 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:42 crc kubenswrapper[4733]: I1204 17:40:42.308577 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:42 crc kubenswrapper[4733]: E1204 17:40:42.308843 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:40:42 crc kubenswrapper[4733]: E1204 17:40:42.308970 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs podName:79d2d07f-2f47-4970-b6e5-7d25b02e639e nodeName:}" failed. No retries permitted until 2025-12-04 17:41:46.308936983 +0000 UTC m=+168.264298059 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs") pod "network-metrics-daemon-mlwj8" (UID: "79d2d07f-2f47-4970-b6e5-7d25b02e639e") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 17:40:42 crc kubenswrapper[4733]: I1204 17:40:42.335605 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:42 crc kubenswrapper[4733]: I1204 17:40:42.335660 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:42 crc kubenswrapper[4733]: E1204 17:40:42.335867 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:42 crc kubenswrapper[4733]: E1204 17:40:42.336782 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:43 crc kubenswrapper[4733]: I1204 17:40:43.335044 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:43 crc kubenswrapper[4733]: I1204 17:40:43.335090 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:43 crc kubenswrapper[4733]: E1204 17:40:43.335233 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:43 crc kubenswrapper[4733]: E1204 17:40:43.335366 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:44 crc kubenswrapper[4733]: I1204 17:40:44.334411 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:44 crc kubenswrapper[4733]: I1204 17:40:44.334502 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:44 crc kubenswrapper[4733]: E1204 17:40:44.334596 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:44 crc kubenswrapper[4733]: E1204 17:40:44.334667 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:45 crc kubenswrapper[4733]: I1204 17:40:45.334559 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:45 crc kubenswrapper[4733]: I1204 17:40:45.334814 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:45 crc kubenswrapper[4733]: E1204 17:40:45.334999 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:45 crc kubenswrapper[4733]: E1204 17:40:45.335224 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:46 crc kubenswrapper[4733]: I1204 17:40:46.334367 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:46 crc kubenswrapper[4733]: I1204 17:40:46.334444 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:46 crc kubenswrapper[4733]: E1204 17:40:46.334533 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:46 crc kubenswrapper[4733]: E1204 17:40:46.334608 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:47 crc kubenswrapper[4733]: I1204 17:40:47.335185 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:47 crc kubenswrapper[4733]: I1204 17:40:47.335218 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:47 crc kubenswrapper[4733]: E1204 17:40:47.335858 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:47 crc kubenswrapper[4733]: E1204 17:40:47.335969 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:47 crc kubenswrapper[4733]: I1204 17:40:47.336347 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:40:47 crc kubenswrapper[4733]: E1204 17:40:47.336615 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qp9sx_openshift-ovn-kubernetes(067d57d5-c6ec-463d-b1a0-ec5e8c686453)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" Dec 04 17:40:48 crc kubenswrapper[4733]: I1204 17:40:48.335716 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:48 crc kubenswrapper[4733]: E1204 17:40:48.337745 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:48 crc kubenswrapper[4733]: I1204 17:40:48.337790 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:48 crc kubenswrapper[4733]: E1204 17:40:48.338186 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:49 crc kubenswrapper[4733]: I1204 17:40:49.334701 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:49 crc kubenswrapper[4733]: I1204 17:40:49.334839 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:49 crc kubenswrapper[4733]: E1204 17:40:49.334917 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:49 crc kubenswrapper[4733]: E1204 17:40:49.334996 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:50 crc kubenswrapper[4733]: I1204 17:40:50.335466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:50 crc kubenswrapper[4733]: I1204 17:40:50.335492 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:50 crc kubenswrapper[4733]: E1204 17:40:50.335875 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:50 crc kubenswrapper[4733]: E1204 17:40:50.336026 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:51 crc kubenswrapper[4733]: I1204 17:40:51.335015 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:51 crc kubenswrapper[4733]: I1204 17:40:51.335174 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:51 crc kubenswrapper[4733]: E1204 17:40:51.335401 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:51 crc kubenswrapper[4733]: E1204 17:40:51.335962 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:52 crc kubenswrapper[4733]: I1204 17:40:52.335060 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:52 crc kubenswrapper[4733]: I1204 17:40:52.335239 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:52 crc kubenswrapper[4733]: E1204 17:40:52.335418 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:52 crc kubenswrapper[4733]: E1204 17:40:52.335572 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:53 crc kubenswrapper[4733]: I1204 17:40:53.335122 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:53 crc kubenswrapper[4733]: I1204 17:40:53.335145 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:53 crc kubenswrapper[4733]: E1204 17:40:53.335416 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:53 crc kubenswrapper[4733]: E1204 17:40:53.335555 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:54 crc kubenswrapper[4733]: I1204 17:40:54.335293 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:54 crc kubenswrapper[4733]: E1204 17:40:54.335426 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:54 crc kubenswrapper[4733]: I1204 17:40:54.335312 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:54 crc kubenswrapper[4733]: E1204 17:40:54.335589 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:55 crc kubenswrapper[4733]: I1204 17:40:55.335106 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:55 crc kubenswrapper[4733]: I1204 17:40:55.335225 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:55 crc kubenswrapper[4733]: E1204 17:40:55.335285 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:55 crc kubenswrapper[4733]: E1204 17:40:55.335427 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.335044 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.335179 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:56 crc kubenswrapper[4733]: E1204 17:40:56.335246 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:56 crc kubenswrapper[4733]: E1204 17:40:56.335392 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.602267 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/1.log" Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.603208 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/0.log" Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.603334 4733 generic.go:334] "Generic (PLEG): container finished" podID="829fda7a-6a96-4007-b2f5-0be03782a764" containerID="dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886" exitCode=1 Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.603407 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-472m8" event={"ID":"829fda7a-6a96-4007-b2f5-0be03782a764","Type":"ContainerDied","Data":"dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886"} Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.603483 4733 scope.go:117] "RemoveContainer" containerID="8ec8802b1d9751830190077e77dcf55931ebf338ba0000b4870d827a449657e6" Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.604030 4733 scope.go:117] "RemoveContainer" containerID="dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886" Dec 04 17:40:56 crc kubenswrapper[4733]: E1204 17:40:56.604262 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-472m8_openshift-multus(829fda7a-6a96-4007-b2f5-0be03782a764)\"" pod="openshift-multus/multus-472m8" podUID="829fda7a-6a96-4007-b2f5-0be03782a764" Dec 04 17:40:56 crc kubenswrapper[4733]: I1204 17:40:56.630781 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9mln9" podStartSLOduration=93.630740173 podStartE2EDuration="1m33.630740173s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:40:32.534368007 +0000 UTC m=+94.489729073" watchObservedRunningTime="2025-12-04 17:40:56.630740173 +0000 UTC m=+118.586101229" Dec 04 17:40:57 crc kubenswrapper[4733]: I1204 17:40:57.334925 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:57 crc kubenswrapper[4733]: E1204 17:40:57.335059 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:40:57 crc kubenswrapper[4733]: I1204 17:40:57.334927 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:57 crc kubenswrapper[4733]: E1204 17:40:57.335135 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:57 crc kubenswrapper[4733]: I1204 17:40:57.609773 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/1.log" Dec 04 17:40:58 crc kubenswrapper[4733]: E1204 17:40:58.317071 4733 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 04 17:40:58 crc kubenswrapper[4733]: I1204 17:40:58.335058 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:40:58 crc kubenswrapper[4733]: I1204 17:40:58.335153 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:40:58 crc kubenswrapper[4733]: E1204 17:40:58.337855 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:40:58 crc kubenswrapper[4733]: E1204 17:40:58.338014 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:40:58 crc kubenswrapper[4733]: E1204 17:40:58.444411 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 17:40:59 crc kubenswrapper[4733]: I1204 17:40:59.335360 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:40:59 crc kubenswrapper[4733]: I1204 17:40:59.335443 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:40:59 crc kubenswrapper[4733]: E1204 17:40:59.335548 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:40:59 crc kubenswrapper[4733]: E1204 17:40:59.335637 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:41:00 crc kubenswrapper[4733]: I1204 17:41:00.334431 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:00 crc kubenswrapper[4733]: I1204 17:41:00.334536 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:00 crc kubenswrapper[4733]: E1204 17:41:00.334629 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:41:00 crc kubenswrapper[4733]: E1204 17:41:00.334741 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:41:01 crc kubenswrapper[4733]: I1204 17:41:01.334683 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:01 crc kubenswrapper[4733]: I1204 17:41:01.334772 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:01 crc kubenswrapper[4733]: E1204 17:41:01.334876 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:41:01 crc kubenswrapper[4733]: E1204 17:41:01.335253 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:41:01 crc kubenswrapper[4733]: I1204 17:41:01.335539 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:41:01 crc kubenswrapper[4733]: I1204 17:41:01.626337 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/3.log" Dec 04 17:41:01 crc kubenswrapper[4733]: I1204 17:41:01.629779 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerStarted","Data":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} Dec 04 17:41:01 crc kubenswrapper[4733]: I1204 17:41:01.634105 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:41:01 crc kubenswrapper[4733]: I1204 17:41:01.671653 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podStartSLOduration=98.671604824 podStartE2EDuration="1m38.671604824s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:01.670970667 +0000 UTC m=+123.626331723" watchObservedRunningTime="2025-12-04 17:41:01.671604824 +0000 UTC m=+123.626965910" Dec 04 17:41:02 crc kubenswrapper[4733]: I1204 17:41:02.259095 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mlwj8"] Dec 04 17:41:02 crc kubenswrapper[4733]: I1204 17:41:02.259280 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:02 crc kubenswrapper[4733]: E1204 17:41:02.259464 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:41:02 crc kubenswrapper[4733]: I1204 17:41:02.334815 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:02 crc kubenswrapper[4733]: E1204 17:41:02.334968 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:41:03 crc kubenswrapper[4733]: I1204 17:41:03.334853 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:03 crc kubenswrapper[4733]: I1204 17:41:03.334888 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:03 crc kubenswrapper[4733]: E1204 17:41:03.335483 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:41:03 crc kubenswrapper[4733]: E1204 17:41:03.335561 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:41:03 crc kubenswrapper[4733]: E1204 17:41:03.446425 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 17:41:04 crc kubenswrapper[4733]: I1204 17:41:04.335188 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:04 crc kubenswrapper[4733]: I1204 17:41:04.335302 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:04 crc kubenswrapper[4733]: E1204 17:41:04.335397 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:41:04 crc kubenswrapper[4733]: E1204 17:41:04.335503 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:41:05 crc kubenswrapper[4733]: I1204 17:41:05.335320 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:05 crc kubenswrapper[4733]: I1204 17:41:05.335334 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:05 crc kubenswrapper[4733]: E1204 17:41:05.335472 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:41:05 crc kubenswrapper[4733]: E1204 17:41:05.335583 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:41:06 crc kubenswrapper[4733]: I1204 17:41:06.334569 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:06 crc kubenswrapper[4733]: E1204 17:41:06.334833 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:41:06 crc kubenswrapper[4733]: I1204 17:41:06.335231 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:06 crc kubenswrapper[4733]: E1204 17:41:06.335379 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:41:07 crc kubenswrapper[4733]: I1204 17:41:07.334664 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:07 crc kubenswrapper[4733]: I1204 17:41:07.334790 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:07 crc kubenswrapper[4733]: E1204 17:41:07.334882 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:41:07 crc kubenswrapper[4733]: E1204 17:41:07.335017 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:41:08 crc kubenswrapper[4733]: I1204 17:41:08.335480 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:08 crc kubenswrapper[4733]: I1204 17:41:08.335631 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:08 crc kubenswrapper[4733]: E1204 17:41:08.338234 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:41:08 crc kubenswrapper[4733]: E1204 17:41:08.338509 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:41:08 crc kubenswrapper[4733]: E1204 17:41:08.447074 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 17:41:09 crc kubenswrapper[4733]: I1204 17:41:09.335290 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:09 crc kubenswrapper[4733]: E1204 17:41:09.335485 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:41:09 crc kubenswrapper[4733]: I1204 17:41:09.335294 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:09 crc kubenswrapper[4733]: E1204 17:41:09.336078 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:41:10 crc kubenswrapper[4733]: I1204 17:41:10.334467 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:10 crc kubenswrapper[4733]: I1204 17:41:10.334492 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:10 crc kubenswrapper[4733]: E1204 17:41:10.334736 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:41:10 crc kubenswrapper[4733]: E1204 17:41:10.334789 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:41:11 crc kubenswrapper[4733]: I1204 17:41:11.337064 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:11 crc kubenswrapper[4733]: I1204 17:41:11.337195 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:11 crc kubenswrapper[4733]: E1204 17:41:11.337481 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:41:11 crc kubenswrapper[4733]: E1204 17:41:11.338286 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:41:11 crc kubenswrapper[4733]: I1204 17:41:11.340645 4733 scope.go:117] "RemoveContainer" containerID="dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886" Dec 04 17:41:11 crc kubenswrapper[4733]: I1204 17:41:11.674147 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/1.log" Dec 04 17:41:11 crc kubenswrapper[4733]: I1204 17:41:11.674222 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-472m8" event={"ID":"829fda7a-6a96-4007-b2f5-0be03782a764","Type":"ContainerStarted","Data":"05f1cdc7acc55339ce7a4767f128c0cfbb18eb7c3649d4d49c2ac1680976c7bc"} Dec 04 17:41:12 crc kubenswrapper[4733]: I1204 17:41:12.335151 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:12 crc kubenswrapper[4733]: I1204 17:41:12.335182 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:12 crc kubenswrapper[4733]: E1204 17:41:12.335353 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 17:41:12 crc kubenswrapper[4733]: E1204 17:41:12.335516 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mlwj8" podUID="79d2d07f-2f47-4970-b6e5-7d25b02e639e" Dec 04 17:41:13 crc kubenswrapper[4733]: I1204 17:41:13.335158 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:13 crc kubenswrapper[4733]: I1204 17:41:13.335246 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:13 crc kubenswrapper[4733]: E1204 17:41:13.335366 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 17:41:13 crc kubenswrapper[4733]: E1204 17:41:13.335544 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 17:41:14 crc kubenswrapper[4733]: I1204 17:41:14.334993 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:14 crc kubenswrapper[4733]: I1204 17:41:14.335039 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:14 crc kubenswrapper[4733]: I1204 17:41:14.338022 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 17:41:14 crc kubenswrapper[4733]: I1204 17:41:14.338048 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 17:41:14 crc kubenswrapper[4733]: I1204 17:41:14.339148 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 17:41:14 crc kubenswrapper[4733]: I1204 17:41:14.339377 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 17:41:15 crc kubenswrapper[4733]: I1204 17:41:15.334542 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:15 crc kubenswrapper[4733]: I1204 17:41:15.334557 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:15 crc kubenswrapper[4733]: I1204 17:41:15.336563 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 17:41:15 crc kubenswrapper[4733]: I1204 17:41:15.337134 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.549960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.608032 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.608566 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.610188 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8j2hf"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.610971 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.615272 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.616066 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.616724 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.617894 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rs6c6"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.618120 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.618318 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.618468 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.620074 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.629870 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.632562 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w9vhc"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.639273 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.641722 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.643407 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.644981 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.644139 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.644699 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.647290 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.647512 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.648009 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.653221 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.655230 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.655616 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.666104 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.666279 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.666435 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.666761 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-rkdfs"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.666944 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.667101 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qjjzl"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.667320 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.667533 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.667942 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.668267 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.667536 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.668787 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.669361 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.670061 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.673495 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.676258 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.676475 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.677285 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.678596 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.682046 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.683241 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.683950 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5fzhh"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.684505 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rhg5t"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.685000 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.685855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.686428 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-87dsb"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.687379 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.691444 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.691461 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.692366 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.692516 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.692655 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.692886 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.694659 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-2b26n"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.695435 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2b26n" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.699363 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.699603 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.700124 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.700366 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.701050 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.701281 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.701662 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.701768 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.701884 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.701994 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.702069 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.702218 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.702852 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.703337 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.703751 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.706366 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.707873 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.708419 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2gzm8"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.708676 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.709410 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.715557 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.716378 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.716867 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.717235 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.729599 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.737481 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.739437 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.739822 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.741764 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.742032 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.744651 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.746162 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.746494 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.746902 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.747071 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.747137 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.747347 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.747598 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.748348 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.748528 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.748391 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.748893 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.749229 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.749423 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.749592 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.749740 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.750127 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.750373 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.763963 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764049 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764101 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764223 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764334 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764466 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764554 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764582 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764660 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764687 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.764971 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765121 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765208 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765300 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765409 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765500 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765649 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765726 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765840 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765914 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.765776 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766093 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766166 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766189 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766278 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766285 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766378 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766405 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766058 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.766534 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.767334 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.767562 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.768359 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.768381 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.771114 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.771212 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.771523 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.772588 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.774647 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.775447 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.775967 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.776016 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.776123 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.776221 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.776362 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.777846 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.778680 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780216 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780242 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780261 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-etcd-client\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780276 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-serving-cert\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780391 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c26584cc-3f5a-47ca-91ae-4f457eb7947e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780408 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-serving-cert\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780423 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-encryption-config\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780438 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-serving-cert\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780463 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-oauth-config\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780478 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-etcd-client\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780495 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv6zb\" (UniqueName: \"kubernetes.io/projected/89610ab1-de41-4937-96d6-47c62e241281-kube-api-access-xv6zb\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780513 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-config\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780526 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-service-ca\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780540 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-encryption-config\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780626 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-trusted-ca\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780693 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-trusted-ca-bundle\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780713 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-config\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780779 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e4a43e7-58da-4eb1-930f-1350cde7398b-serving-cert\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780851 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-config\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780920 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780939 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26584cc-3f5a-47ca-91ae-4f457eb7947e-config\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780955 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-image-import-ca\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780981 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-client-ca\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.780995 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c26584cc-3f5a-47ca-91ae-4f457eb7947e-images\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781012 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kjgr\" (UniqueName: \"kubernetes.io/projected/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-kube-api-access-8kjgr\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781030 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-serving-cert\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781045 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781064 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781112 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-audit\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781162 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-client-ca\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781183 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6q7d\" (UniqueName: \"kubernetes.io/projected/6c99fe8f-586c-4e35-b010-73dad59c59af-kube-api-access-j6q7d\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781202 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6zt7\" (UniqueName: \"kubernetes.io/projected/c26584cc-3f5a-47ca-91ae-4f457eb7947e-kube-api-access-r6zt7\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781224 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-config\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781252 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnjvr\" (UniqueName: \"kubernetes.io/projected/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-kube-api-access-xnjvr\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781275 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8lgt\" (UniqueName: \"kubernetes.io/projected/e3966d4d-94a4-4059-a535-fe2eb759c1cf-kube-api-access-b8lgt\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781297 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-console-config\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781317 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781334 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3966d4d-94a4-4059-a535-fe2eb759c1cf-audit-dir\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781364 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-serving-cert\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781383 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781402 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3966d4d-94a4-4059-a535-fe2eb759c1cf-node-pullsecrets\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781427 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgk9"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781752 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.782762 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.781425 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-etcd-serving-ca\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.783588 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nb862"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.784416 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.784830 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.785302 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.785431 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89610ab1-de41-4937-96d6-47c62e241281-audit-dir\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.785575 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95nrs\" (UniqueName: \"kubernetes.io/projected/8e4a43e7-58da-4eb1-930f-1350cde7398b-kube-api-access-95nrs\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.785638 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-oauth-serving-cert\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.785671 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-audit-policies\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.785721 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgckf\" (UniqueName: \"kubernetes.io/projected/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-kube-api-access-xgckf\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.786261 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.786702 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.786770 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.787151 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.787566 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.787928 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.788674 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.789268 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.790620 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bnzzn"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.791684 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v9mkr"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.791909 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.792008 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.793054 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.793927 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-872sg"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.794597 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.795274 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7qgps"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.796281 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.796574 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.797008 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.798211 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qjjzl"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.803119 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8j2hf"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.803689 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rkdfs"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.810667 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.816400 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5fzhh"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.818486 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2gzm8"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.820087 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rhg5t"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.821223 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.822859 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.823214 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kpfdb"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.824869 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w9vhc"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.824977 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kpfdb" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.825403 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.826890 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nb862"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.828249 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.828976 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.830479 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.830497 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2b26n"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.832317 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rs6c6"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.833287 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.834279 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.835442 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.836260 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.837288 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgk9"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.838253 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.839290 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.840251 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n6k4k"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.841918 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rzbdr"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.842172 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.843052 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.843109 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.843533 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.844209 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.845170 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.846156 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.847198 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.848175 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.849145 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-872sg"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.850097 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rzbdr"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.850425 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.851101 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v9mkr"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.852073 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n6k4k"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.853058 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7qgps"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.854028 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.855093 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.856130 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bnzzn"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.857125 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kpfdb"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.858162 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ct5h2"] Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.858751 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.870579 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.887748 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26584cc-3f5a-47ca-91ae-4f457eb7947e-config\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.887814 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-image-import-ca\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.887865 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-client-ca\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.887887 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c26584cc-3f5a-47ca-91ae-4f457eb7947e-images\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.888950 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kjgr\" (UniqueName: \"kubernetes.io/projected/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-kube-api-access-8kjgr\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.888999 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-serving-cert\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889323 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26584cc-3f5a-47ca-91ae-4f457eb7947e-config\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889221 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889475 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c26584cc-3f5a-47ca-91ae-4f457eb7947e-images\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889518 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-audit\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889713 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqshg\" (UniqueName: \"kubernetes.io/projected/64240304-b67a-4285-8832-a531f608eae4-kube-api-access-jqshg\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-client-ca\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889826 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6q7d\" (UniqueName: \"kubernetes.io/projected/6c99fe8f-586c-4e35-b010-73dad59c59af-kube-api-access-j6q7d\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889863 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6zt7\" (UniqueName: \"kubernetes.io/projected/c26584cc-3f5a-47ca-91ae-4f457eb7947e-kube-api-access-r6zt7\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnjvr\" (UniqueName: \"kubernetes.io/projected/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-kube-api-access-xnjvr\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889762 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-client-ca\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889930 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.889929 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-config\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890020 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8lgt\" (UniqueName: \"kubernetes.io/projected/e3966d4d-94a4-4059-a535-fe2eb759c1cf-kube-api-access-b8lgt\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890049 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64240304-b67a-4285-8832-a531f608eae4-config\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890075 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-serving-cert\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890098 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-console-config\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890141 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3966d4d-94a4-4059-a535-fe2eb759c1cf-audit-dir\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890175 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890195 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3966d4d-94a4-4059-a535-fe2eb759c1cf-node-pullsecrets\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890214 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-etcd-serving-ca\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890238 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89610ab1-de41-4937-96d6-47c62e241281-audit-dir\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890270 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgckf\" (UniqueName: \"kubernetes.io/projected/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-kube-api-access-xgckf\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890293 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95nrs\" (UniqueName: \"kubernetes.io/projected/8e4a43e7-58da-4eb1-930f-1350cde7398b-kube-api-access-95nrs\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890317 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-oauth-serving-cert\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890334 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-audit-policies\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890361 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890410 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890432 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-etcd-client\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890458 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-serving-cert\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890487 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-encryption-config\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890518 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c26584cc-3f5a-47ca-91ae-4f457eb7947e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890562 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-serving-cert\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890599 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-serving-cert\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890589 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-image-import-ca\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890651 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-oauth-config\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890671 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-etcd-client\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890739 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-client-ca\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890754 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890739 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv6zb\" (UniqueName: \"kubernetes.io/projected/89610ab1-de41-4937-96d6-47c62e241281-kube-api-access-xv6zb\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890825 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-service-ca\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890850 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-config\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890870 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-encryption-config\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-trusted-ca\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890921 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/64240304-b67a-4285-8832-a531f608eae4-machine-approver-tls\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890948 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/64240304-b67a-4285-8832-a531f608eae4-auth-proxy-config\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890971 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-trusted-ca-bundle\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.890998 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e4a43e7-58da-4eb1-930f-1350cde7398b-serving-cert\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.891021 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-config\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.891045 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-config\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.891065 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.891197 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-config\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.891756 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-audit\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.892273 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-config\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.893307 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89610ab1-de41-4937-96d6-47c62e241281-audit-dir\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.894070 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-config\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.894348 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-service-ca\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.895036 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.895463 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-config\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.895478 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-serving-cert\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.895834 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-trusted-ca\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896036 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-oauth-serving-cert\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896252 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-serving-cert\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896257 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-etcd-client\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896326 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-audit-policies\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896780 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896614 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896832 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3966d4d-94a4-4059-a535-fe2eb759c1cf-node-pullsecrets\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896422 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-serving-cert\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.896647 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3966d4d-94a4-4059-a535-fe2eb759c1cf-audit-dir\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.897295 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/89610ab1-de41-4937-96d6-47c62e241281-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.897316 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-serving-cert\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.897532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e3966d4d-94a4-4059-a535-fe2eb759c1cf-etcd-serving-ca\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.898853 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/89610ab1-de41-4937-96d6-47c62e241281-encryption-config\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.899023 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-etcd-client\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.899329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-trusted-ca-bundle\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.899460 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.899588 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-console-config\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.900024 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e3966d4d-94a4-4059-a535-fe2eb759c1cf-encryption-config\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.900430 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e4a43e7-58da-4eb1-930f-1350cde7398b-serving-cert\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.900483 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c26584cc-3f5a-47ca-91ae-4f457eb7947e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.901306 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-oauth-config\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.903926 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-serving-cert\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.911649 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.930621 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.952827 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.970579 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.993678 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.994087 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/64240304-b67a-4285-8832-a531f608eae4-machine-approver-tls\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.994204 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/64240304-b67a-4285-8832-a531f608eae4-auth-proxy-config\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.994349 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqshg\" (UniqueName: \"kubernetes.io/projected/64240304-b67a-4285-8832-a531f608eae4-kube-api-access-jqshg\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.994488 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64240304-b67a-4285-8832-a531f608eae4-config\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.994808 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/64240304-b67a-4285-8832-a531f608eae4-auth-proxy-config\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.994971 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64240304-b67a-4285-8832-a531f608eae4-config\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:22 crc kubenswrapper[4733]: I1204 17:41:22.997745 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/64240304-b67a-4285-8832-a531f608eae4-machine-approver-tls\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.030815 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.051949 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.070601 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.090580 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.111326 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.131003 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.151777 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.171591 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.191594 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.211847 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.231691 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.251533 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.270981 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.291123 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.311042 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.331685 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.350753 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.380050 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.391682 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.411214 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.431975 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.450976 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.470730 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.492331 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.511628 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.532034 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.551500 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.591693 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.611544 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.632481 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.651395 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.671994 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.691450 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.711371 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.732616 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.752107 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.772762 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.800770 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.809315 4733 request.go:700] Waited for 1.017048827s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/secrets?fieldSelector=metadata.name%3Dv4-0-config-system-ocp-branding-template&limit=500&resourceVersion=0 Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.831490 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.832232 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.854278 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.871198 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.891282 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.920417 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.930686 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.951497 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.971258 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 17:41:23 crc kubenswrapper[4733]: I1204 17:41:23.990875 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.010473 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.031296 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.051978 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.071610 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.092553 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.122823 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.131140 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.151728 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.171028 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.191530 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.212459 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.231595 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.250830 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.271747 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.291923 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.312131 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.331656 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.353173 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.371381 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.392505 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.411499 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.432071 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.451469 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.471996 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.490470 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.510519 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.531413 4733 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.551393 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.571201 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.591713 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.611400 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.631317 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.650872 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.670934 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.717752 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kjgr\" (UniqueName: \"kubernetes.io/projected/d37c4b9d-058b-4f74-9987-e4cc9fa80db2-kube-api-access-8kjgr\") pod \"openshift-apiserver-operator-796bbdcf4f-f95hh\" (UID: \"d37c4b9d-058b-4f74-9987-e4cc9fa80db2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.751069 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6zt7\" (UniqueName: \"kubernetes.io/projected/c26584cc-3f5a-47ca-91ae-4f457eb7947e-kube-api-access-r6zt7\") pod \"machine-api-operator-5694c8668f-w9vhc\" (UID: \"c26584cc-3f5a-47ca-91ae-4f457eb7947e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.753545 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.765208 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6q7d\" (UniqueName: \"kubernetes.io/projected/6c99fe8f-586c-4e35-b010-73dad59c59af-kube-api-access-j6q7d\") pod \"console-f9d7485db-rkdfs\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.781920 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv6zb\" (UniqueName: \"kubernetes.io/projected/89610ab1-de41-4937-96d6-47c62e241281-kube-api-access-xv6zb\") pod \"apiserver-7bbb656c7d-w57sg\" (UID: \"89610ab1-de41-4937-96d6-47c62e241281\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.803110 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8lgt\" (UniqueName: \"kubernetes.io/projected/e3966d4d-94a4-4059-a535-fe2eb759c1cf-kube-api-access-b8lgt\") pod \"apiserver-76f77b778f-rs6c6\" (UID: \"e3966d4d-94a4-4059-a535-fe2eb759c1cf\") " pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.804303 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.821772 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgckf\" (UniqueName: \"kubernetes.io/projected/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-kube-api-access-xgckf\") pod \"controller-manager-879f6c89f-8j2hf\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.829566 4733 request.go:700] Waited for 1.93364651s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.838123 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95nrs\" (UniqueName: \"kubernetes.io/projected/8e4a43e7-58da-4eb1-930f-1350cde7398b-kube-api-access-95nrs\") pod \"route-controller-manager-6576b87f9c-vbf92\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.848430 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.861117 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnjvr\" (UniqueName: \"kubernetes.io/projected/cb2a9fa8-747f-4fe9-ac33-043df6745c1b-kube-api-access-xnjvr\") pod \"console-operator-58897d9998-qjjzl\" (UID: \"cb2a9fa8-747f-4fe9-ac33-043df6745c1b\") " pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.870805 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.879994 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.881538 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqshg\" (UniqueName: \"kubernetes.io/projected/64240304-b67a-4285-8832-a531f608eae4-kube-api-access-jqshg\") pod \"machine-approver-56656f9798-v7nq9\" (UID: \"64240304-b67a-4285-8832-a531f608eae4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.887399 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.908171 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.925527 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.927142 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgf6k\" (UniqueName: \"kubernetes.io/projected/315bbafb-4080-4284-9d82-5f0f202e5c11-kube-api-access-kgf6k\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.927274 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-tls\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.927331 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2cc37bbd-c101-44f0-a327-d6dc3012e375-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.927634 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/70ad32ad-a01a-4676-96f0-399bb8afcbfd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.927679 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23c1be98-c468-4a48-9d55-dfaeefef722c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.927705 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpvrb\" (UniqueName: \"kubernetes.io/projected/86e1c39b-aa71-43dc-b331-a66850269dc0-kube-api-access-lpvrb\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.927930 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-bound-sa-token\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.927970 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5ee72b2-5573-41a9-9c53-8cd27878b57c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928029 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e612af-44a0-40fe-aeb4-143d2e19c51d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928144 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c670b992-b3cc-4495-ad15-2b1ec738b3a6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vghxl\" (UID: \"c670b992-b3cc-4495-ad15-2b1ec738b3a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928342 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1c147058-60bd-403a-b3c6-47750b8e40a6-srv-cert\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928387 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xflss\" (UniqueName: \"kubernetes.io/projected/c5ee72b2-5573-41a9-9c53-8cd27878b57c-kube-api-access-xflss\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928425 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77rsq\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-kube-api-access-77rsq\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928450 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/660082e5-0ecd-44ab-b9ac-1672eae922d6-config\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928636 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928682 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5ee72b2-5573-41a9-9c53-8cd27878b57c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928706 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-default-certificate\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928728 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cc37bbd-c101-44f0-a327-d6dc3012e375-config\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928752 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zhjh\" (UniqueName: \"kubernetes.io/projected/4f8163e4-ba70-4cb2-b719-a52e1801ab74-kube-api-access-9zhjh\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.928867 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:24 crc kubenswrapper[4733]: E1204 17:41:24.929154 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.429140486 +0000 UTC m=+147.384501642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929183 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86e1c39b-aa71-43dc-b331-a66850269dc0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929213 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/315bbafb-4080-4284-9d82-5f0f202e5c11-proxy-tls\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929235 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cc37bbd-c101-44f0-a327-d6dc3012e375-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929261 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5ee72b2-5573-41a9-9c53-8cd27878b57c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929315 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/70ad32ad-a01a-4676-96f0-399bb8afcbfd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929390 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-metrics-certs\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929416 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4f8163e4-ba70-4cb2-b719-a52e1801ab74-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929669 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfv97\" (UniqueName: \"kubernetes.io/projected/c670b992-b3cc-4495-ad15-2b1ec738b3a6-kube-api-access-mfv97\") pod \"cluster-samples-operator-665b6dd947-vghxl\" (UID: \"c670b992-b3cc-4495-ad15-2b1ec738b3a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929701 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/315bbafb-4080-4284-9d82-5f0f202e5c11-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929738 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-certificates\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929764 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929806 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lh5l\" (UniqueName: \"kubernetes.io/projected/35292c5a-4dcc-44d6-8c3b-c1640378a894-kube-api-access-6lh5l\") pod \"dns-operator-744455d44c-2gzm8\" (UID: \"35292c5a-4dcc-44d6-8c3b-c1640378a894\") " pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929959 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkql8\" (UniqueName: \"kubernetes.io/projected/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-kube-api-access-fkql8\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.929988 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn4ck\" (UniqueName: \"kubernetes.io/projected/06e612af-44a0-40fe-aeb4-143d2e19c51d-kube-api-access-wn4ck\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930067 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/35292c5a-4dcc-44d6-8c3b-c1640378a894-metrics-tls\") pod \"dns-operator-744455d44c-2gzm8\" (UID: \"35292c5a-4dcc-44d6-8c3b-c1640378a894\") " pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930095 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85ndd\" (UniqueName: \"kubernetes.io/projected/4a92a269-34fd-4e5a-88b2-9a020e651909-kube-api-access-85ndd\") pod \"migrator-59844c95c7-mcr28\" (UID: \"4a92a269-34fd-4e5a-88b2-9a020e651909\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930118 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ppp8\" (UniqueName: \"kubernetes.io/projected/23c1be98-c468-4a48-9d55-dfaeefef722c-kube-api-access-7ppp8\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930139 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660082e5-0ecd-44ab-b9ac-1672eae922d6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930464 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/624f7b69-b973-4294-8043-8747bf3d256f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nb862\" (UID: \"624f7b69-b973-4294-8043-8747bf3d256f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930503 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9trk2\" (UniqueName: \"kubernetes.io/projected/8e123c23-4206-4c39-8549-9cf6727aff8c-kube-api-access-9trk2\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930529 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e123c23-4206-4c39-8549-9cf6727aff8c-service-ca-bundle\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930551 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86e1c39b-aa71-43dc-b331-a66850269dc0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930602 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7db6v\" (UniqueName: \"kubernetes.io/projected/1c147058-60bd-403a-b3c6-47750b8e40a6-kube-api-access-7db6v\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930637 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-trusted-ca\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930662 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e612af-44a0-40fe-aeb4-143d2e19c51d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930687 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgmvl\" (UniqueName: \"kubernetes.io/projected/624f7b69-b973-4294-8043-8747bf3d256f-kube-api-access-wgmvl\") pod \"multus-admission-controller-857f4d67dd-nb862\" (UID: \"624f7b69-b973-4294-8043-8747bf3d256f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930715 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23c1be98-c468-4a48-9d55-dfaeefef722c-serving-cert\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930734 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1c147058-60bd-403a-b3c6-47750b8e40a6-profile-collector-cert\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930756 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4f8163e4-ba70-4cb2-b719-a52e1801ab74-proxy-tls\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930804 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-stats-auth\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930828 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gslq2\" (UniqueName: \"kubernetes.io/projected/70ad32ad-a01a-4676-96f0-399bb8afcbfd-kube-api-access-gslq2\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930915 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70ad32ad-a01a-4676-96f0-399bb8afcbfd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930958 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.930979 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/660082e5-0ecd-44ab-b9ac-1672eae922d6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.931004 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/315bbafb-4080-4284-9d82-5f0f202e5c11-images\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:24 crc kubenswrapper[4733]: I1204 17:41:24.951402 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.032940 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033429 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/70ad32ad-a01a-4676-96f0-399bb8afcbfd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033455 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpvrb\" (UniqueName: \"kubernetes.io/projected/86e1c39b-aa71-43dc-b331-a66850269dc0-kube-api-access-lpvrb\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033472 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23c1be98-c468-4a48-9d55-dfaeefef722c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033493 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-bound-sa-token\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033512 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97tb8\" (UniqueName: \"kubernetes.io/projected/20a96bf5-566b-4da4-8869-0d27c4b9ef3c-kube-api-access-97tb8\") pod \"downloads-7954f5f757-2b26n\" (UID: \"20a96bf5-566b-4da4-8869-0d27c4b9ef3c\") " pod="openshift-console/downloads-7954f5f757-2b26n" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033527 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-policies\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033543 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033560 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkr9k\" (UniqueName: \"kubernetes.io/projected/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-kube-api-access-gkr9k\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033577 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/aadecb45-b22c-4202-b37b-ba3d071883e9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-csj6s\" (UID: \"aadecb45-b22c-4202-b37b-ba3d071883e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033596 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-service-ca-bundle\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033618 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-serving-cert\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033633 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-service-ca\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033650 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77rsq\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-kube-api-access-77rsq\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033671 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5ee72b2-5573-41a9-9c53-8cd27878b57c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033690 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zhjh\" (UniqueName: \"kubernetes.io/projected/4f8163e4-ba70-4cb2-b719-a52e1801ab74-kube-api-access-9zhjh\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033707 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlg9d\" (UniqueName: \"kubernetes.io/projected/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-kube-api-access-hlg9d\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033723 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hmbj\" (UniqueName: \"kubernetes.io/projected/aadecb45-b22c-4202-b37b-ba3d071883e9-kube-api-access-4hmbj\") pod \"package-server-manager-789f6589d5-csj6s\" (UID: \"aadecb45-b22c-4202-b37b-ba3d071883e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033739 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033755 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgbk2\" (UniqueName: \"kubernetes.io/projected/e6f87509-6a44-4fbd-85f8-b579f3bff55a-kube-api-access-qgbk2\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033774 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cc37bbd-c101-44f0-a327-d6dc3012e375-config\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033815 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033833 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/315bbafb-4080-4284-9d82-5f0f202e5c11-proxy-tls\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033857 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cc37bbd-c101-44f0-a327-d6dc3012e375-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033874 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5ee72b2-5573-41a9-9c53-8cd27878b57c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033890 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/70ad32ad-a01a-4676-96f0-399bb8afcbfd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033906 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6rxk\" (UniqueName: \"kubernetes.io/projected/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-kube-api-access-d6rxk\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033923 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-metrics-certs\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033938 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4f8163e4-ba70-4cb2-b719-a52e1801ab74-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033958 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033974 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzxkg\" (UniqueName: \"kubernetes.io/projected/741e419c-3e6a-4e3b-80b0-f48900f974f1-kube-api-access-tzxkg\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.033988 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-csi-data-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034006 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/315bbafb-4080-4284-9d82-5f0f202e5c11-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034022 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034040 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034055 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9540e997-8c29-41ba-a743-93c1c62c5fcb-apiservice-cert\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034071 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-config\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034089 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkql8\" (UniqueName: \"kubernetes.io/projected/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-kube-api-access-fkql8\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034104 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-config-volume\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034117 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d3ee16-a81a-414f-86a7-77f1ebfa2d40-cert\") pod \"ingress-canary-kpfdb\" (UID: \"86d3ee16-a81a-414f-86a7-77f1ebfa2d40\") " pod="openshift-ingress-canary/ingress-canary-kpfdb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034133 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/35292c5a-4dcc-44d6-8c3b-c1640378a894-metrics-tls\") pod \"dns-operator-744455d44c-2gzm8\" (UID: \"35292c5a-4dcc-44d6-8c3b-c1640378a894\") " pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034149 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034167 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660082e5-0ecd-44ab-b9ac-1672eae922d6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034184 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7pzn\" (UniqueName: \"kubernetes.io/projected/86d3ee16-a81a-414f-86a7-77f1ebfa2d40-kube-api-access-k7pzn\") pod \"ingress-canary-kpfdb\" (UID: \"86d3ee16-a81a-414f-86a7-77f1ebfa2d40\") " pod="openshift-ingress-canary/ingress-canary-kpfdb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034201 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td76w\" (UniqueName: \"kubernetes.io/projected/b156d126-a527-43ad-a025-cac37696669a-kube-api-access-td76w\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034224 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034246 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e123c23-4206-4c39-8549-9cf6727aff8c-service-ca-bundle\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034266 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86e1c39b-aa71-43dc-b331-a66850269dc0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034282 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034300 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/741e419c-3e6a-4e3b-80b0-f48900f974f1-node-bootstrap-token\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034316 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-trusted-ca\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034332 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e612af-44a0-40fe-aeb4-143d2e19c51d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034348 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9540e997-8c29-41ba-a743-93c1c62c5fcb-webhook-cert\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034363 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1c147058-60bd-403a-b3c6-47750b8e40a6-profile-collector-cert\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70ad32ad-a01a-4676-96f0-399bb8afcbfd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034413 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-mountpoint-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034432 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034448 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-signing-key\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034466 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/315bbafb-4080-4284-9d82-5f0f202e5c11-images\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034487 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034503 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgf6k\" (UniqueName: \"kubernetes.io/projected/315bbafb-4080-4284-9d82-5f0f202e5c11-kube-api-access-kgf6k\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034520 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2cc37bbd-c101-44f0-a327-d6dc3012e375-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5ee72b2-5573-41a9-9c53-8cd27878b57c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034554 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e612af-44a0-40fe-aeb4-143d2e19c51d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034569 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c670b992-b3cc-4495-ad15-2b1ec738b3a6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vghxl\" (UID: \"c670b992-b3cc-4495-ad15-2b1ec738b3a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034586 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-serving-cert\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034601 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dj4m\" (UniqueName: \"kubernetes.io/projected/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-kube-api-access-4dj4m\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034616 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-socket-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034633 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1c147058-60bd-403a-b3c6-47750b8e40a6-srv-cert\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034648 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xflss\" (UniqueName: \"kubernetes.io/projected/c5ee72b2-5573-41a9-9c53-8cd27878b57c-kube-api-access-xflss\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034665 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/660082e5-0ecd-44ab-b9ac-1672eae922d6-config\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034679 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/741e419c-3e6a-4e3b-80b0-f48900f974f1-certs\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034696 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-default-certificate\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034739 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034755 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86e1c39b-aa71-43dc-b331-a66850269dc0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034770 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-ca\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034785 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b156d126-a527-43ad-a025-cac37696669a-config-volume\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034815 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-plugins-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034831 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034850 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfv97\" (UniqueName: \"kubernetes.io/projected/c670b992-b3cc-4495-ad15-2b1ec738b3a6-kube-api-access-mfv97\") pod \"cluster-samples-operator-665b6dd947-vghxl\" (UID: \"c670b992-b3cc-4495-ad15-2b1ec738b3a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034867 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lh5l\" (UniqueName: \"kubernetes.io/projected/35292c5a-4dcc-44d6-8c3b-c1640378a894-kube-api-access-6lh5l\") pod \"dns-operator-744455d44c-2gzm8\" (UID: \"35292c5a-4dcc-44d6-8c3b-c1640378a894\") " pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034883 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-certificates\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034899 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034915 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhzrl\" (UniqueName: \"kubernetes.io/projected/ae851a3f-c43e-462e-b083-f8d2ba558f55-kube-api-access-lhzrl\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034929 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-signing-cabundle\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034952 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-config\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034969 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7767c7f9-42fd-4525-903b-3e6073dd694f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-v7zc2\" (UID: \"7767c7f9-42fd-4525-903b-3e6073dd694f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.034989 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85ndd\" (UniqueName: \"kubernetes.io/projected/4a92a269-34fd-4e5a-88b2-9a020e651909-kube-api-access-85ndd\") pod \"migrator-59844c95c7-mcr28\" (UID: \"4a92a269-34fd-4e5a-88b2-9a020e651909\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035007 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn4ck\" (UniqueName: \"kubernetes.io/projected/06e612af-44a0-40fe-aeb4-143d2e19c51d-kube-api-access-wn4ck\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035022 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ppp8\" (UniqueName: \"kubernetes.io/projected/23c1be98-c468-4a48-9d55-dfaeefef722c-kube-api-access-7ppp8\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035041 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/624f7b69-b973-4294-8043-8747bf3d256f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nb862\" (UID: \"624f7b69-b973-4294-8043-8747bf3d256f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035056 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9trk2\" (UniqueName: \"kubernetes.io/projected/8e123c23-4206-4c39-8549-9cf6727aff8c-kube-api-access-9trk2\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035072 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-config\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035089 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035104 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7db6v\" (UniqueName: \"kubernetes.io/projected/1c147058-60bd-403a-b3c6-47750b8e40a6-kube-api-access-7db6v\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035119 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-dir\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035110 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23c1be98-c468-4a48-9d55-dfaeefef722c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035136 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035320 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-srv-cert\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035381 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w22sp\" (UniqueName: \"kubernetes.io/projected/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-kube-api-access-w22sp\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035429 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9lnb\" (UniqueName: \"kubernetes.io/projected/7767c7f9-42fd-4525-903b-3e6073dd694f-kube-api-access-f9lnb\") pod \"control-plane-machine-set-operator-78cbb6b69f-v7zc2\" (UID: \"7767c7f9-42fd-4525-903b-3e6073dd694f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035468 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgmvl\" (UniqueName: \"kubernetes.io/projected/624f7b69-b973-4294-8043-8747bf3d256f-kube-api-access-wgmvl\") pod \"multus-admission-controller-857f4d67dd-nb862\" (UID: \"624f7b69-b973-4294-8043-8747bf3d256f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035517 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23c1be98-c468-4a48-9d55-dfaeefef722c-serving-cert\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035552 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xpvp\" (UniqueName: \"kubernetes.io/projected/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-kube-api-access-8xpvp\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035679 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035701 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-secret-volume\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035741 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4f8163e4-ba70-4cb2-b719-a52e1801ab74-proxy-tls\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035763 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-stats-auth\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035780 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gslq2\" (UniqueName: \"kubernetes.io/projected/70ad32ad-a01a-4676-96f0-399bb8afcbfd-kube-api-access-gslq2\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035831 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-serving-cert\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035903 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-client\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035928 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mqfj\" (UniqueName: \"kubernetes.io/projected/9540e997-8c29-41ba-a743-93c1c62c5fcb-kube-api-access-8mqfj\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035948 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/660082e5-0ecd-44ab-b9ac-1672eae922d6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.035986 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9540e997-8c29-41ba-a743-93c1c62c5fcb-tmpfs\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.036004 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b156d126-a527-43ad-a025-cac37696669a-metrics-tls\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.036024 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-tls\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.036065 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.036083 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-registration-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.037912 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cc37bbd-c101-44f0-a327-d6dc3012e375-config\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.040719 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5ee72b2-5573-41a9-9c53-8cd27878b57c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.042396 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/315bbafb-4080-4284-9d82-5f0f202e5c11-images\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.043379 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/35292c5a-4dcc-44d6-8c3b-c1640378a894-metrics-tls\") pod \"dns-operator-744455d44c-2gzm8\" (UID: \"35292c5a-4dcc-44d6-8c3b-c1640378a894\") " pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.043585 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-stats-auth\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.043758 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/315bbafb-4080-4284-9d82-5f0f202e5c11-proxy-tls\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.044147 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.544109747 +0000 UTC m=+147.499470793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.044450 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23c1be98-c468-4a48-9d55-dfaeefef722c-serving-cert\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.045176 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.045475 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/70ad32ad-a01a-4676-96f0-399bb8afcbfd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.045713 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-certificates\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.052631 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e123c23-4206-4c39-8549-9cf6727aff8c-service-ca-bundle\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.055002 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.055424 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/660082e5-0ecd-44ab-b9ac-1672eae922d6-config\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.055997 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.056124 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5ee72b2-5573-41a9-9c53-8cd27878b57c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.056239 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e612af-44a0-40fe-aeb4-143d2e19c51d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.057184 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70ad32ad-a01a-4676-96f0-399bb8afcbfd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.058829 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/660082e5-0ecd-44ab-b9ac-1672eae922d6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.062458 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/315bbafb-4080-4284-9d82-5f0f202e5c11-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.062480 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.064174 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1c147058-60bd-403a-b3c6-47750b8e40a6-srv-cert\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.064190 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/624f7b69-b973-4294-8043-8747bf3d256f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nb862\" (UID: \"624f7b69-b973-4294-8043-8747bf3d256f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.065028 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1c147058-60bd-403a-b3c6-47750b8e40a6-profile-collector-cert\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.066181 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-default-certificate\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.068304 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e612af-44a0-40fe-aeb4-143d2e19c51d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.068683 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e123c23-4206-4c39-8549-9cf6727aff8c-metrics-certs\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.070138 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cc37bbd-c101-44f0-a327-d6dc3012e375-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.070546 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4f8163e4-ba70-4cb2-b719-a52e1801ab74-proxy-tls\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.070559 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86e1c39b-aa71-43dc-b331-a66850269dc0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.071268 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4f8163e4-ba70-4cb2-b719-a52e1801ab74-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.079566 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-tls\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.080204 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-trusted-ca\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.080754 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86e1c39b-aa71-43dc-b331-a66850269dc0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.081409 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c670b992-b3cc-4495-ad15-2b1ec738b3a6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vghxl\" (UID: \"c670b992-b3cc-4495-ad15-2b1ec738b3a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.081640 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-bound-sa-token\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.085507 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpvrb\" (UniqueName: \"kubernetes.io/projected/86e1c39b-aa71-43dc-b331-a66850269dc0-kube-api-access-lpvrb\") pod \"kube-storage-version-migrator-operator-b67b599dd-mmssx\" (UID: \"86e1c39b-aa71-43dc-b331-a66850269dc0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.089759 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.103705 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77rsq\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-kube-api-access-77rsq\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.128371 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgmvl\" (UniqueName: \"kubernetes.io/projected/624f7b69-b973-4294-8043-8747bf3d256f-kube-api-access-wgmvl\") pod \"multus-admission-controller-857f4d67dd-nb862\" (UID: \"624f7b69-b973-4294-8043-8747bf3d256f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.136943 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.136991 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgbk2\" (UniqueName: \"kubernetes.io/projected/e6f87509-6a44-4fbd-85f8-b579f3bff55a-kube-api-access-qgbk2\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137026 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlg9d\" (UniqueName: \"kubernetes.io/projected/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-kube-api-access-hlg9d\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137057 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hmbj\" (UniqueName: \"kubernetes.io/projected/aadecb45-b22c-4202-b37b-ba3d071883e9-kube-api-access-4hmbj\") pod \"package-server-manager-789f6589d5-csj6s\" (UID: \"aadecb45-b22c-4202-b37b-ba3d071883e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137081 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137113 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6rxk\" (UniqueName: \"kubernetes.io/projected/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-kube-api-access-d6rxk\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137146 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137172 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzxkg\" (UniqueName: \"kubernetes.io/projected/741e419c-3e6a-4e3b-80b0-f48900f974f1-kube-api-access-tzxkg\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137195 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137216 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-csi-data-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137248 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137273 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9540e997-8c29-41ba-a743-93c1c62c5fcb-apiservice-cert\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137295 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-config\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137320 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-config-volume\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137343 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d3ee16-a81a-414f-86a7-77f1ebfa2d40-cert\") pod \"ingress-canary-kpfdb\" (UID: \"86d3ee16-a81a-414f-86a7-77f1ebfa2d40\") " pod="openshift-ingress-canary/ingress-canary-kpfdb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137373 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7pzn\" (UniqueName: \"kubernetes.io/projected/86d3ee16-a81a-414f-86a7-77f1ebfa2d40-kube-api-access-k7pzn\") pod \"ingress-canary-kpfdb\" (UID: \"86d3ee16-a81a-414f-86a7-77f1ebfa2d40\") " pod="openshift-ingress-canary/ingress-canary-kpfdb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137399 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137425 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137446 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td76w\" (UniqueName: \"kubernetes.io/projected/b156d126-a527-43ad-a025-cac37696669a-kube-api-access-td76w\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137468 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/741e419c-3e6a-4e3b-80b0-f48900f974f1-node-bootstrap-token\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137494 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137517 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9540e997-8c29-41ba-a743-93c1c62c5fcb-webhook-cert\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137554 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137576 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-mountpoint-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137598 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-signing-key\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137660 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-serving-cert\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137684 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dj4m\" (UniqueName: \"kubernetes.io/projected/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-kube-api-access-4dj4m\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137719 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-socket-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137742 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/741e419c-3e6a-4e3b-80b0-f48900f974f1-certs\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137774 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137816 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137840 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-ca\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b156d126-a527-43ad-a025-cac37696669a-config-volume\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137884 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137912 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-plugins-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137953 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhzrl\" (UniqueName: \"kubernetes.io/projected/ae851a3f-c43e-462e-b083-f8d2ba558f55-kube-api-access-lhzrl\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.137977 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-signing-cabundle\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138008 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-config\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138056 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7767c7f9-42fd-4525-903b-3e6073dd694f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-v7zc2\" (UID: \"7767c7f9-42fd-4525-903b-3e6073dd694f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138101 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-config\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138124 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138152 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w22sp\" (UniqueName: \"kubernetes.io/projected/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-kube-api-access-w22sp\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138173 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9lnb\" (UniqueName: \"kubernetes.io/projected/7767c7f9-42fd-4525-903b-3e6073dd694f-kube-api-access-f9lnb\") pod \"control-plane-machine-set-operator-78cbb6b69f-v7zc2\" (UID: \"7767c7f9-42fd-4525-903b-3e6073dd694f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138189 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-dir\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138206 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138223 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-srv-cert\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138237 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xpvp\" (UniqueName: \"kubernetes.io/projected/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-kube-api-access-8xpvp\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138253 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138270 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-secret-volume\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138299 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-serving-cert\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138316 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-client\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138335 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mqfj\" (UniqueName: \"kubernetes.io/projected/9540e997-8c29-41ba-a743-93c1c62c5fcb-kube-api-access-8mqfj\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138351 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9540e997-8c29-41ba-a743-93c1c62c5fcb-tmpfs\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138367 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b156d126-a527-43ad-a025-cac37696669a-metrics-tls\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-registration-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138412 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138433 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138454 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97tb8\" (UniqueName: \"kubernetes.io/projected/20a96bf5-566b-4da4-8869-0d27c4b9ef3c-kube-api-access-97tb8\") pod \"downloads-7954f5f757-2b26n\" (UID: \"20a96bf5-566b-4da4-8869-0d27c4b9ef3c\") " pod="openshift-console/downloads-7954f5f757-2b26n" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138474 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-policies\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138494 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkr9k\" (UniqueName: \"kubernetes.io/projected/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-kube-api-access-gkr9k\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138525 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/aadecb45-b22c-4202-b37b-ba3d071883e9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-csj6s\" (UID: \"aadecb45-b22c-4202-b37b-ba3d071883e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138548 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-service-ca-bundle\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138571 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-service-ca\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138592 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-serving-cert\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.138717 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.139308 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-csi-data-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.139750 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qjjzl"] Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.139810 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rs6c6"] Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.140350 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-config\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.140522 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9540e997-8c29-41ba-a743-93c1c62c5fcb-tmpfs\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.140589 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-config-volume\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.141005 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-dir\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.141761 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-registration-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.141906 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.142051 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-mountpoint-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.142344 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9540e997-8c29-41ba-a743-93c1c62c5fcb-apiservice-cert\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.142755 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.142891 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-policies\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.143022 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.143234 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b156d126-a527-43ad-a025-cac37696669a-config-volume\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.144166 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.144613 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-service-ca\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.144755 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.144773 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-service-ca-bundle\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.144869 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-socket-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.145447 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-signing-cabundle\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.145861 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.145913 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e6f87509-6a44-4fbd-85f8-b579f3bff55a-plugins-dir\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.146192 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.646177772 +0000 UTC m=+147.601538818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.146501 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-ca\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.148035 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9540e997-8c29-41ba-a743-93c1c62c5fcb-webhook-cert\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.148824 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-config\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.149390 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-config\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.150432 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7767c7f9-42fd-4525-903b-3e6073dd694f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-v7zc2\" (UID: \"7767c7f9-42fd-4525-903b-3e6073dd694f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.150921 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.151434 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.152349 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.154553 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-serving-cert\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.156784 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.159341 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-serving-cert\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.159668 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zhjh\" (UniqueName: \"kubernetes.io/projected/4f8163e4-ba70-4cb2-b719-a52e1801ab74-kube-api-access-9zhjh\") pod \"machine-config-controller-84d6567774-7rff2\" (UID: \"4f8163e4-ba70-4cb2-b719-a52e1801ab74\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.159692 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-serving-cert\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.160140 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b156d126-a527-43ad-a025-cac37696669a-metrics-tls\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.160172 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-secret-volume\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.160199 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.160350 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/741e419c-3e6a-4e3b-80b0-f48900f974f1-certs\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.160458 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.160572 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.160873 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86d3ee16-a81a-414f-86a7-77f1ebfa2d40-cert\") pod \"ingress-canary-kpfdb\" (UID: \"86d3ee16-a81a-414f-86a7-77f1ebfa2d40\") " pod="openshift-ingress-canary/ingress-canary-kpfdb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.162184 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-srv-cert\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.163397 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/741e419c-3e6a-4e3b-80b0-f48900f974f1-node-bootstrap-token\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.163926 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.164783 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-etcd-client\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.164831 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/aadecb45-b22c-4202-b37b-ba3d071883e9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-csj6s\" (UID: \"aadecb45-b22c-4202-b37b-ba3d071883e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.165032 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-signing-key\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.188967 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660082e5-0ecd-44ab-b9ac-1672eae922d6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mq9bc\" (UID: \"660082e5-0ecd-44ab-b9ac-1672eae922d6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.205036 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gslq2\" (UniqueName: \"kubernetes.io/projected/70ad32ad-a01a-4676-96f0-399bb8afcbfd-kube-api-access-gslq2\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.224670 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgf6k\" (UniqueName: \"kubernetes.io/projected/315bbafb-4080-4284-9d82-5f0f202e5c11-kube-api-access-kgf6k\") pod \"machine-config-operator-74547568cd-tdkmq\" (UID: \"315bbafb-4080-4284-9d82-5f0f202e5c11\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.239483 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.240152 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.740134558 +0000 UTC m=+147.695495604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.242537 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh"] Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.244847 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2cc37bbd-c101-44f0-a327-d6dc3012e375-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zzvkk\" (UID: \"2cc37bbd-c101-44f0-a327-d6dc3012e375\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:25 crc kubenswrapper[4733]: W1204 17:41:25.250272 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd37c4b9d_058b_4f74_9987_e4cc9fa80db2.slice/crio-b9e1d1cacb47073d13a74b85735b4b0e611a86b43dc28c7078cc8cb293d37925 WatchSource:0}: Error finding container b9e1d1cacb47073d13a74b85735b4b0e611a86b43dc28c7078cc8cb293d37925: Status 404 returned error can't find the container with id b9e1d1cacb47073d13a74b85735b4b0e611a86b43dc28c7078cc8cb293d37925 Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.265185 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfv97\" (UniqueName: \"kubernetes.io/projected/c670b992-b3cc-4495-ad15-2b1ec738b3a6-kube-api-access-mfv97\") pod \"cluster-samples-operator-665b6dd947-vghxl\" (UID: \"c670b992-b3cc-4495-ad15-2b1ec738b3a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.275912 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8j2hf"] Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.283726 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lh5l\" (UniqueName: \"kubernetes.io/projected/35292c5a-4dcc-44d6-8c3b-c1640378a894-kube-api-access-6lh5l\") pod \"dns-operator-744455d44c-2gzm8\" (UID: \"35292c5a-4dcc-44d6-8c3b-c1640378a894\") " pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.291130 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.302910 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkql8\" (UniqueName: \"kubernetes.io/projected/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-kube-api-access-fkql8\") pod \"marketplace-operator-79b997595-kzgk9\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.326958 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ppp8\" (UniqueName: \"kubernetes.io/projected/23c1be98-c468-4a48-9d55-dfaeefef722c-kube-api-access-7ppp8\") pod \"openshift-config-operator-7777fb866f-kk5vw\" (UID: \"23c1be98-c468-4a48-9d55-dfaeefef722c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.338238 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.340899 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.341400 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.84137782 +0000 UTC m=+147.796738906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.344916 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.349215 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85ndd\" (UniqueName: \"kubernetes.io/projected/4a92a269-34fd-4e5a-88b2-9a020e651909-kube-api-access-85ndd\") pod \"migrator-59844c95c7-mcr28\" (UID: \"4a92a269-34fd-4e5a-88b2-9a020e651909\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.351629 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.358503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.364252 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.364954 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn4ck\" (UniqueName: \"kubernetes.io/projected/06e612af-44a0-40fe-aeb4-143d2e19c51d-kube-api-access-wn4ck\") pod \"openshift-controller-manager-operator-756b6f6bc6-jz45r\" (UID: \"06e612af-44a0-40fe-aeb4-143d2e19c51d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.372402 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.386472 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.387540 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5ee72b2-5573-41a9-9c53-8cd27878b57c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.396925 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w9vhc"] Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.398020 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rkdfs"] Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.408289 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.411753 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92"] Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.427009 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg"] Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.429287 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xflss\" (UniqueName: \"kubernetes.io/projected/c5ee72b2-5573-41a9-9c53-8cd27878b57c-kube-api-access-xflss\") pod \"ingress-operator-5b745b69d9-5rpdj\" (UID: \"c5ee72b2-5573-41a9-9c53-8cd27878b57c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.429462 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/70ad32ad-a01a-4676-96f0-399bb8afcbfd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5qnjl\" (UID: \"70ad32ad-a01a-4676-96f0-399bb8afcbfd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.441978 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.442220 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.94218019 +0000 UTC m=+147.897541276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.442329 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.442478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.442566 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.442729 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.442876 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.443332 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:25.943311541 +0000 UTC m=+147.898672617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.447869 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.448236 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.449426 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.457346 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9trk2\" (UniqueName: \"kubernetes.io/projected/8e123c23-4206-4c39-8549-9cf6727aff8c-kube-api-access-9trk2\") pod \"router-default-5444994796-87dsb\" (UID: \"8e123c23-4206-4c39-8549-9cf6727aff8c\") " pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.460432 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.465511 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7db6v\" (UniqueName: \"kubernetes.io/projected/1c147058-60bd-403a-b3c6-47750b8e40a6-kube-api-access-7db6v\") pod \"catalog-operator-68c6474976-226s9\" (UID: \"1c147058-60bd-403a-b3c6-47750b8e40a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.489540 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzxkg\" (UniqueName: \"kubernetes.io/projected/741e419c-3e6a-4e3b-80b0-f48900f974f1-kube-api-access-tzxkg\") pod \"machine-config-server-ct5h2\" (UID: \"741e419c-3e6a-4e3b-80b0-f48900f974f1\") " pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.503711 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ct5h2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.507101 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.510491 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-khb78\" (UID: \"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.526022 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.543876 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.544966 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.044930173 +0000 UTC m=+148.000291259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.546100 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6rxk\" (UniqueName: \"kubernetes.io/projected/54a44568-f9d9-41c8-9a5b-ca12e3c3231c-kube-api-access-d6rxk\") pod \"authentication-operator-69f744f599-v9mkr\" (UID: \"54a44568-f9d9-41c8-9a5b-ca12e3c3231c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.556666 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.557608 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.563138 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.569008 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w22sp\" (UniqueName: \"kubernetes.io/projected/fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2-kube-api-access-w22sp\") pod \"etcd-operator-b45778765-5fzhh\" (UID: \"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.570605 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.583817 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.586074 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgbk2\" (UniqueName: \"kubernetes.io/projected/e6f87509-6a44-4fbd-85f8-b579f3bff55a-kube-api-access-qgbk2\") pod \"csi-hostpathplugin-n6k4k\" (UID: \"e6f87509-6a44-4fbd-85f8-b579f3bff55a\") " pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.605710 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlg9d\" (UniqueName: \"kubernetes.io/projected/d11858a1-e21b-4dfd-bfa4-5cde50b14b6a-kube-api-access-hlg9d\") pod \"service-ca-operator-777779d784-7qgps\" (UID: \"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.627567 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.641578 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.642554 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7pzn\" (UniqueName: \"kubernetes.io/projected/86d3ee16-a81a-414f-86a7-77f1ebfa2d40-kube-api-access-k7pzn\") pod \"ingress-canary-kpfdb\" (UID: \"86d3ee16-a81a-414f-86a7-77f1ebfa2d40\") " pod="openshift-ingress-canary/ingress-canary-kpfdb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.647699 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.648214 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.148194579 +0000 UTC m=+148.103555635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.648262 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hmbj\" (UniqueName: \"kubernetes.io/projected/aadecb45-b22c-4202-b37b-ba3d071883e9-kube-api-access-4hmbj\") pod \"package-server-manager-789f6589d5-csj6s\" (UID: \"aadecb45-b22c-4202-b37b-ba3d071883e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:25 crc kubenswrapper[4733]: W1204 17:41:25.656304 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6ba934e_90ea_47d0_9c7f_7ceca992d5b7.slice/crio-cbad3c1658fbbf362bfc71e73e1417d6aabeb88bc2cc80c59bf823f02c83879e WatchSource:0}: Error finding container cbad3c1658fbbf362bfc71e73e1417d6aabeb88bc2cc80c59bf823f02c83879e: Status 404 returned error can't find the container with id cbad3c1658fbbf362bfc71e73e1417d6aabeb88bc2cc80c59bf823f02c83879e Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.659509 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97tb8\" (UniqueName: \"kubernetes.io/projected/20a96bf5-566b-4da4-8869-0d27c4b9ef3c-kube-api-access-97tb8\") pod \"downloads-7954f5f757-2b26n\" (UID: \"20a96bf5-566b-4da4-8869-0d27c4b9ef3c\") " pod="openshift-console/downloads-7954f5f757-2b26n" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.670537 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9lnb\" (UniqueName: \"kubernetes.io/projected/7767c7f9-42fd-4525-903b-3e6073dd694f-kube-api-access-f9lnb\") pod \"control-plane-machine-set-operator-78cbb6b69f-v7zc2\" (UID: \"7767c7f9-42fd-4525-903b-3e6073dd694f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.680753 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.704262 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mqfj\" (UniqueName: \"kubernetes.io/projected/9540e997-8c29-41ba-a743-93c1c62c5fcb-kube-api-access-8mqfj\") pod \"packageserver-d55dfcdfc-9c59g\" (UID: \"9540e997-8c29-41ba-a743-93c1c62c5fcb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.712894 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.718861 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xpvp\" (UniqueName: \"kubernetes.io/projected/ca36c16c-7eba-43dd-a88c-5f9d18b26d00-kube-api-access-8xpvp\") pod \"olm-operator-6b444d44fb-frmrq\" (UID: \"ca36c16c-7eba-43dd-a88c-5f9d18b26d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.726241 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.731851 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td76w\" (UniqueName: \"kubernetes.io/projected/b156d126-a527-43ad-a025-cac37696669a-kube-api-access-td76w\") pod \"dns-default-rzbdr\" (UID: \"b156d126-a527-43ad-a025-cac37696669a\") " pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.740489 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.748311 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.748514 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.248489216 +0000 UTC m=+148.203850252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.748606 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.748763 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkr9k\" (UniqueName: \"kubernetes.io/projected/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-kube-api-access-gkr9k\") pod \"collect-profiles-29414490-dhv68\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.748939 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.248927668 +0000 UTC m=+148.204288704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.751294 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.760239 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" event={"ID":"e3966d4d-94a4-4059-a535-fe2eb759c1cf","Type":"ContainerStarted","Data":"c904130f468226d469e9342412719b574b037fb2a891859fb9b9df5e21fd166a"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.763084 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" event={"ID":"c26584cc-3f5a-47ca-91ae-4f457eb7947e","Type":"ContainerStarted","Data":"4bb45711c24d6e52eadd57c1f719b168a0104b61a6b9361555938b2a853b13d8"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.764455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qjjzl" event={"ID":"cb2a9fa8-747f-4fe9-ac33-043df6745c1b","Type":"ContainerStarted","Data":"356bf974362989eaa2ae262f878e9c5eda493ad928d0ecdd8386655697b68b7c"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.764487 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qjjzl" event={"ID":"cb2a9fa8-747f-4fe9-ac33-043df6745c1b","Type":"ContainerStarted","Data":"37aa89b64ff8119aaa8b9fca05bd086178b6cf89e1368bcea628e0e43776755e"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.764808 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.764964 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dj4m\" (UniqueName: \"kubernetes.io/projected/b55dc242-06d5-43dc-bc4c-b9febdbbd7ae-kube-api-access-4dj4m\") pod \"service-ca-9c57cc56f-872sg\" (UID: \"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.765586 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.766054 4733 patch_prober.go:28] interesting pod/console-operator-58897d9998-qjjzl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.766107 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qjjzl" podUID="cb2a9fa8-747f-4fe9-ac33-043df6745c1b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.766154 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" event={"ID":"64240304-b67a-4285-8832-a531f608eae4","Type":"ContainerStarted","Data":"ada5ffe4b9da4fbb006a66131543810bbe3cd8a62b2b68842fb72e645130af28"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.766183 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" event={"ID":"64240304-b67a-4285-8832-a531f608eae4","Type":"ContainerStarted","Data":"453706d988f5d178b1df69185151fb46e310cc0fb9f9cd6c4ae4751ecbe940d8"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.768167 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" event={"ID":"89610ab1-de41-4937-96d6-47c62e241281","Type":"ContainerStarted","Data":"33758581726c1901eff1f402fd9e19d22111eace8f75812febdb56e5c75bfc03"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.769781 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" event={"ID":"d37c4b9d-058b-4f74-9987-e4cc9fa80db2","Type":"ContainerStarted","Data":"b9e1d1cacb47073d13a74b85735b4b0e611a86b43dc28c7078cc8cb293d37925"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.771475 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" event={"ID":"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7","Type":"ContainerStarted","Data":"cbad3c1658fbbf362bfc71e73e1417d6aabeb88bc2cc80c59bf823f02c83879e"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.772080 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.774286 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rkdfs" event={"ID":"6c99fe8f-586c-4e35-b010-73dad59c59af","Type":"ContainerStarted","Data":"d945ec397b06f1c9f8b95b2d720c7ec23c87b5a85357c8d8fef53ea99fcd8d35"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.776504 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" event={"ID":"8e4a43e7-58da-4eb1-930f-1350cde7398b","Type":"ContainerStarted","Data":"0b08c4d845f60f35971aaf4a7fcd9f73668939933d7b7db9f7f22eaaebb28273"} Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.785409 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhzrl\" (UniqueName: \"kubernetes.io/projected/ae851a3f-c43e-462e-b083-f8d2ba558f55-kube-api-access-lhzrl\") pod \"oauth-openshift-558db77b4-bnzzn\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.791319 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.798430 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.816391 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.839306 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kpfdb" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.845558 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2b26n" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.854214 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.855621 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.855844 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.355814312 +0000 UTC m=+148.311175378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.856843 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.857296 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.357281061 +0000 UTC m=+148.312642107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.958412 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.958965 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.458938244 +0000 UTC m=+148.414299290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:25 crc kubenswrapper[4733]: I1204 17:41:25.959589 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:25 crc kubenswrapper[4733]: E1204 17:41:25.960256 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.460240899 +0000 UTC m=+148.415601945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.019888 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq"] Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.057392 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.057412 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.058962 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-872sg" Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.060151 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.060460 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.560445143 +0000 UTC m=+148.515806189 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.162107 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.162603 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.66256269 +0000 UTC m=+148.617923736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.263673 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.263843 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.763821512 +0000 UTC m=+148.719182558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.265086 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.265562 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.765554318 +0000 UTC m=+148.720915364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.284902 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2gzm8"] Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.366388 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.367081 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.867063688 +0000 UTC m=+148.822424724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.445023 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc"] Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.471089 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.471437 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:26.971422794 +0000 UTC m=+148.926783840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.505343 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qjjzl" podStartSLOduration=123.505327045 podStartE2EDuration="2m3.505327045s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:26.47390076 +0000 UTC m=+148.429261806" watchObservedRunningTime="2025-12-04 17:41:26.505327045 +0000 UTC m=+148.460688091" Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.549131 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nb862"] Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.572730 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.573340 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.073319713 +0000 UTC m=+149.028680759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.578365 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgk9"] Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.673984 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.674246 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.174234926 +0000 UTC m=+149.129595972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: W1204 17:41:26.760004 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod624f7b69_b973_4294_8043_8747bf3d256f.slice/crio-a87b0b185e21ef5be696f72a2e66082129503195abe1acc8ae1373ea46d55c3c WatchSource:0}: Error finding container a87b0b185e21ef5be696f72a2e66082129503195abe1acc8ae1373ea46d55c3c: Status 404 returned error can't find the container with id a87b0b185e21ef5be696f72a2e66082129503195abe1acc8ae1373ea46d55c3c Dec 04 17:41:26 crc kubenswrapper[4733]: W1204 17:41:26.770953 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7caf3411_fca7_4025_b9b6_eb3b4a43ab81.slice/crio-8aeff507b88a4b15055d0b3b7717627cc13d40ef5a348bdc3eb9bec4f5534cd6 WatchSource:0}: Error finding container 8aeff507b88a4b15055d0b3b7717627cc13d40ef5a348bdc3eb9bec4f5534cd6: Status 404 returned error can't find the container with id 8aeff507b88a4b15055d0b3b7717627cc13d40ef5a348bdc3eb9bec4f5534cd6 Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.798497 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.799269 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.299252598 +0000 UTC m=+149.254613644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.839934 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" event={"ID":"315bbafb-4080-4284-9d82-5f0f202e5c11","Type":"ContainerStarted","Data":"c2e801994d576292512d0364ae42a341195b3bce3c810279ab5370c8a1ab2bb5"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.839977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" event={"ID":"315bbafb-4080-4284-9d82-5f0f202e5c11","Type":"ContainerStarted","Data":"bf6d0234cc26a21d191d3cd0a3b5cd57af6735d9c23e9ed0e27c89afd65731cf"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.899956 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" event={"ID":"c26584cc-3f5a-47ca-91ae-4f457eb7947e","Type":"ContainerStarted","Data":"8634a248ab2fdc29a4582107cf19286899c5fd557d8010c336b3199e2bce19b5"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.900670 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:26 crc kubenswrapper[4733]: E1204 17:41:26.902548 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.402537165 +0000 UTC m=+149.357898201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.937172 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" event={"ID":"624f7b69-b973-4294-8043-8747bf3d256f","Type":"ContainerStarted","Data":"a87b0b185e21ef5be696f72a2e66082129503195abe1acc8ae1373ea46d55c3c"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.946858 4733 generic.go:334] "Generic (PLEG): container finished" podID="89610ab1-de41-4937-96d6-47c62e241281" containerID="10f30d91eee3bbbaed526fe29f935be192f8f7e887c987115e3390014bc83174" exitCode=0 Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.947101 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" event={"ID":"89610ab1-de41-4937-96d6-47c62e241281","Type":"ContainerDied","Data":"10f30d91eee3bbbaed526fe29f935be192f8f7e887c987115e3390014bc83174"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.954099 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" event={"ID":"d37c4b9d-058b-4f74-9987-e4cc9fa80db2","Type":"ContainerStarted","Data":"a8bab34bee856bb747ae3c4ad4969c2aa1cfb086899a1499e9a0d2bd1fcf1069"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.957337 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-87dsb" event={"ID":"8e123c23-4206-4c39-8549-9cf6727aff8c","Type":"ContainerStarted","Data":"20dae6ff58a18fb0f5ccda29b4b131d89a4e430214990f465e47445f5d7cbc0f"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.957365 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-87dsb" event={"ID":"8e123c23-4206-4c39-8549-9cf6727aff8c","Type":"ContainerStarted","Data":"69c0bab5fd32e42515134631415b8e8015d7c20e543503661707e7fddea77c34"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.970979 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rkdfs" event={"ID":"6c99fe8f-586c-4e35-b010-73dad59c59af","Type":"ContainerStarted","Data":"2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.974281 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ct5h2" event={"ID":"741e419c-3e6a-4e3b-80b0-f48900f974f1","Type":"ContainerStarted","Data":"b5791239b427cae3daf39795fbc1e8bf5290a2c63348891e87c2bcfbdf600fcf"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.974318 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ct5h2" event={"ID":"741e419c-3e6a-4e3b-80b0-f48900f974f1","Type":"ContainerStarted","Data":"253d5565cb26e3b2df4665875cd535b6aca80571aaa0ab6ed50f4bfeaa1ec918"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.980125 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" event={"ID":"8e4a43e7-58da-4eb1-930f-1350cde7398b","Type":"ContainerStarted","Data":"a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.980313 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.982323 4733 generic.go:334] "Generic (PLEG): container finished" podID="e3966d4d-94a4-4059-a535-fe2eb759c1cf" containerID="ad1c898b8588c837766d2f718061a11908b18d6900c0779888a97b7419aeb892" exitCode=0 Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.982352 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" event={"ID":"e3966d4d-94a4-4059-a535-fe2eb759c1cf","Type":"ContainerDied","Data":"ad1c898b8588c837766d2f718061a11908b18d6900c0779888a97b7419aeb892"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.983198 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" event={"ID":"7caf3411-fca7-4025-b9b6-eb3b4a43ab81","Type":"ContainerStarted","Data":"8aeff507b88a4b15055d0b3b7717627cc13d40ef5a348bdc3eb9bec4f5534cd6"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.984434 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" event={"ID":"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7","Type":"ContainerStarted","Data":"ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.984648 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.985568 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" event={"ID":"660082e5-0ecd-44ab-b9ac-1672eae922d6","Type":"ContainerStarted","Data":"2e373a3067f7bfceb1877284c36941c8b6e362b37a42c918bbc4fe35d2875952"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.992919 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" event={"ID":"64240304-b67a-4285-8832-a531f608eae4","Type":"ContainerStarted","Data":"bdf76857bca77fdb91d2728e10084fba1bb6b6f5d73ddcf33f3d57ee7ec47195"} Dec 04 17:41:26 crc kubenswrapper[4733]: I1204 17:41:26.996737 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" event={"ID":"35292c5a-4dcc-44d6-8c3b-c1640378a894","Type":"ContainerStarted","Data":"141554eb7a51df5eed9ce3ce3388fccf7263e6c93b489f4a2af8e0ba87751471"} Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.008562 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.009369 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.012952 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.512932493 +0000 UTC m=+149.468293539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.017532 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qjjzl" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.114564 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.115050 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.615038049 +0000 UTC m=+149.570399095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.219936 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.220961 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.720946046 +0000 UTC m=+149.676307092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.321680 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.322010 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.821998962 +0000 UTC m=+149.777360008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.330574 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-f95hh" podStartSLOduration=124.330555583 podStartE2EDuration="2m4.330555583s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:27.286579321 +0000 UTC m=+149.241940377" watchObservedRunningTime="2025-12-04 17:41:27.330555583 +0000 UTC m=+149.285916629" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.381816 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" podStartSLOduration=123.38178067 podStartE2EDuration="2m3.38178067s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:27.375009418 +0000 UTC m=+149.330370474" watchObservedRunningTime="2025-12-04 17:41:27.38178067 +0000 UTC m=+149.337141706" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.422672 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.423222 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:27.923206714 +0000 UTC m=+149.878567760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.440968 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-rkdfs" podStartSLOduration=124.440952791 podStartE2EDuration="2m4.440952791s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:27.414087499 +0000 UTC m=+149.369448545" watchObservedRunningTime="2025-12-04 17:41:27.440952791 +0000 UTC m=+149.396313837" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.443007 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v7nq9" podStartSLOduration=124.443000407 podStartE2EDuration="2m4.443000407s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:27.441414924 +0000 UTC m=+149.396775970" watchObservedRunningTime="2025-12-04 17:41:27.443000407 +0000 UTC m=+149.398361453" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.486975 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" podStartSLOduration=124.486958158 podStartE2EDuration="2m4.486958158s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:27.483812573 +0000 UTC m=+149.439173619" watchObservedRunningTime="2025-12-04 17:41:27.486958158 +0000 UTC m=+149.442319204" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.525963 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.526299 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.026287765 +0000 UTC m=+149.981648811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.527030 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ct5h2" podStartSLOduration=5.527014615 podStartE2EDuration="5.527014615s" podCreationTimestamp="2025-12-04 17:41:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:27.525361951 +0000 UTC m=+149.480722997" watchObservedRunningTime="2025-12-04 17:41:27.527014615 +0000 UTC m=+149.482375661" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.529538 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.626428 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.626870 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.126847789 +0000 UTC m=+150.082208835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.640746 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:27 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:27 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:27 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.647154 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.651523 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.689779 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-87dsb" podStartSLOduration=124.689761901 podStartE2EDuration="2m4.689761901s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:27.665729555 +0000 UTC m=+149.621090621" watchObservedRunningTime="2025-12-04 17:41:27.689761901 +0000 UTC m=+149.645122937" Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.730106 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.732032 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.232012177 +0000 UTC m=+150.187373223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.839930 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.840327 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.340310908 +0000 UTC m=+150.295671954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.896713 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28"] Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.930341 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw"] Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.942314 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk"] Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.943140 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:27 crc kubenswrapper[4733]: E1204 17:41:27.943908 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.443895333 +0000 UTC m=+150.399256379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.973613 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj"] Dec 04 17:41:27 crc kubenswrapper[4733]: I1204 17:41:27.996272 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.002089 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.037069 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" event={"ID":"c26584cc-3f5a-47ca-91ae-4f457eb7947e","Type":"ContainerStarted","Data":"f072ddd19b22905e88393de2cf01df250c1bc4cfff072ad824eb4b29432360bf"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.044255 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.045151 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.545136326 +0000 UTC m=+150.500497372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.053005 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.067698 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-w9vhc" podStartSLOduration=124.067683032 podStartE2EDuration="2m4.067683032s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:28.06648411 +0000 UTC m=+150.021845146" watchObservedRunningTime="2025-12-04 17:41:28.067683032 +0000 UTC m=+150.023044079" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.075039 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" event={"ID":"624f7b69-b973-4294-8043-8747bf3d256f","Type":"ContainerStarted","Data":"55b1c1b3e84035ec9cae171b760ea8871ee743287b9c91319dd390574a640b4e"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.109007 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.122544 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" event={"ID":"660082e5-0ecd-44ab-b9ac-1672eae922d6","Type":"ContainerStarted","Data":"77da2c6e9b5131ca02c294c84c7c73deb99a8e8abc05fa473ef91e917c66a1b9"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.154630 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.155342 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.655325799 +0000 UTC m=+150.610686845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.181077 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" event={"ID":"315bbafb-4080-4284-9d82-5f0f202e5c11","Type":"ContainerStarted","Data":"bfa0d9408413117ea3504dfff475171d1597dc1111efabf84a7af6b729f74378"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.205946 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" event={"ID":"7caf3411-fca7-4025-b9b6-eb3b4a43ab81","Type":"ContainerStarted","Data":"5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.206680 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.221183 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tdkmq" podStartSLOduration=125.221168998 podStartE2EDuration="2m5.221168998s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:28.220124721 +0000 UTC m=+150.175485767" watchObservedRunningTime="2025-12-04 17:41:28.221168998 +0000 UTC m=+150.176530044" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.221889 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" event={"ID":"23c1be98-c468-4a48-9d55-dfaeefef722c","Type":"ContainerStarted","Data":"446cfd50ac94c6e439a0e8362e6dace77e3b38a8576f904f62a9da24e291f2e0"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.222018 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mq9bc" podStartSLOduration=125.222013272 podStartE2EDuration="2m5.222013272s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:28.154044634 +0000 UTC m=+150.109405680" watchObservedRunningTime="2025-12-04 17:41:28.222013272 +0000 UTC m=+150.177374318" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.240377 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kzgk9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.240423 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.240825 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" event={"ID":"35292c5a-4dcc-44d6-8c3b-c1640378a894","Type":"ContainerStarted","Data":"618b3d53fa43265a74d68c8df6d68c374c8aa2cf510da6ce033efd695b36399e"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.243392 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n6k4k"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.255377 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.255643 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.755628176 +0000 UTC m=+150.710989222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.258501 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" podStartSLOduration=124.258488753 podStartE2EDuration="2m4.258488753s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:28.258083732 +0000 UTC m=+150.213444778" watchObservedRunningTime="2025-12-04 17:41:28.258488753 +0000 UTC m=+150.213849799" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.279466 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.288848 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7qgps"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.298302 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" podStartSLOduration=125.298287102 podStartE2EDuration="2m5.298287102s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:28.292087926 +0000 UTC m=+150.247448972" watchObservedRunningTime="2025-12-04 17:41:28.298287102 +0000 UTC m=+150.253648148" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.290813 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" event={"ID":"89610ab1-de41-4937-96d6-47c62e241281","Type":"ContainerStarted","Data":"0cd69de93a212d1886f688bf7b656bfee00ae0b868f447923ac593c83dee7b53"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.316260 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" event={"ID":"4a92a269-34fd-4e5a-88b2-9a020e651909","Type":"ContainerStarted","Data":"e828d3e4ad455e82f2efd6381d7e2bf8458e03a0e8bdc6bbbb3499e39ad9c80e"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.324878 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" podStartSLOduration=124.324858326 podStartE2EDuration="2m4.324858326s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:28.316236535 +0000 UTC m=+150.271597581" watchObservedRunningTime="2025-12-04 17:41:28.324858326 +0000 UTC m=+150.280219372" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.327236 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kpfdb"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.370312 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.374603 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.874588244 +0000 UTC m=+150.829949290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.411336 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" event={"ID":"e3966d4d-94a4-4059-a535-fe2eb759c1cf","Type":"ContainerStarted","Data":"7d4de9c4235e08ac63ceab31d973363df367921833e4ad8d0c17befd8e987cf5"} Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.411373 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.415945 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.470055 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.471309 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.471756 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.971739966 +0000 UTC m=+150.927101012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.472034 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.473672 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:28.973641888 +0000 UTC m=+150.929002934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.504972 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.510088 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2b26n"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.537388 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bnzzn"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.545056 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:28 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:28 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:28 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.545108 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.549003 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-872sg"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.562771 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.587637 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.587960 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.08794197 +0000 UTC m=+151.043303016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.588157 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.594883 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v9mkr"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.613873 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5fzhh"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.646920 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.658429 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rzbdr"] Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.689161 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.689429 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.189418029 +0000 UTC m=+151.144779075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.792027 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.792924 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.292897611 +0000 UTC m=+151.248258657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:28 crc kubenswrapper[4733]: I1204 17:41:28.894282 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:28 crc kubenswrapper[4733]: E1204 17:41:28.894568 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.394556154 +0000 UTC m=+151.349917200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.001186 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.001750 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.501735096 +0000 UTC m=+151.457096142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.102899 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.103159 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.603149532 +0000 UTC m=+151.558510568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.203970 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.204234 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.70422055 +0000 UTC m=+151.659581596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.305340 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.305669 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.805655818 +0000 UTC m=+151.761016864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.409161 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.409521 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:29.90950427 +0000 UTC m=+151.864865316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.425222 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kpfdb" event={"ID":"86d3ee16-a81a-414f-86a7-77f1ebfa2d40","Type":"ContainerStarted","Data":"98542107930f582d6d0fe5b3ae0cfbd5c5de09a26a3bc1cad8fd9465f7daace7"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.430718 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" event={"ID":"c5ee72b2-5573-41a9-9c53-8cd27878b57c","Type":"ContainerStarted","Data":"97f72a8b9d8ad13d5ebc7e0f539a8a6d71eee25330dd186780fcd059902f6031"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.430765 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" event={"ID":"c5ee72b2-5573-41a9-9c53-8cd27878b57c","Type":"ContainerStarted","Data":"fd045692e3a9bfc56753084d23726b9e9fd376ee7b3620f100af2f19655039d2"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.436175 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-872sg" event={"ID":"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae","Type":"ContainerStarted","Data":"fb0ef256d74a0f0af2548653138aa62ba4fb7f0d4e930394e1d915798e7636bc"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.446408 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" event={"ID":"aadecb45-b22c-4202-b37b-ba3d071883e9","Type":"ContainerStarted","Data":"c623fafd0d318aab9cd375ed46bb9aff3677025a6193490e40b01845c5ed37df"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.451055 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rzbdr" event={"ID":"b156d126-a527-43ad-a025-cac37696669a","Type":"ContainerStarted","Data":"f043cfa134f9b7ccb17067c2aea800cc00b16a0504949a5922f364f8d8428c60"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.484046 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" event={"ID":"ae851a3f-c43e-462e-b083-f8d2ba558f55","Type":"ContainerStarted","Data":"d2b5de879c003f8ebbbb647156997fd37dcbe4469b9ce0c8b9bc137f5bb5e383"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.490728 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" event={"ID":"54a44568-f9d9-41c8-9a5b-ca12e3c3231c","Type":"ContainerStarted","Data":"4a00a08755a6970a4cd61dd6c641d7b34e5339e8aa7679b27ed15cb055ddbbb2"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.490764 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" event={"ID":"54a44568-f9d9-41c8-9a5b-ca12e3c3231c","Type":"ContainerStarted","Data":"2a558ac7ca66849c5cab214cc0ad0eab63de50aa2f23535bbae2593c1b2f7541"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.495528 4733 generic.go:334] "Generic (PLEG): container finished" podID="23c1be98-c468-4a48-9d55-dfaeefef722c" containerID="5fe83856cc6bbabe7d605bcfdd3395679df4cad05969e64ffe4ff8c74f582e27" exitCode=0 Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.495600 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" event={"ID":"23c1be98-c468-4a48-9d55-dfaeefef722c","Type":"ContainerDied","Data":"5fe83856cc6bbabe7d605bcfdd3395679df4cad05969e64ffe4ff8c74f582e27"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.502168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" event={"ID":"4f8163e4-ba70-4cb2-b719-a52e1801ab74","Type":"ContainerStarted","Data":"70e3650006417b41f72f3ad418984c4c5cd1517476ff0cbcac37f9eb13c3a3ef"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.502226 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" event={"ID":"4f8163e4-ba70-4cb2-b719-a52e1801ab74","Type":"ContainerStarted","Data":"2c762d65c38d16485a0c719cc773098254c9906c4bb9694a165363bbea3a46c8"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.503825 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" event={"ID":"ca36c16c-7eba-43dd-a88c-5f9d18b26d00","Type":"ContainerStarted","Data":"5f5e53aecdbf1b15c024194090b7ef328e159f35c735c37b1458bd0b97cb385a"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.503855 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" event={"ID":"ca36c16c-7eba-43dd-a88c-5f9d18b26d00","Type":"ContainerStarted","Data":"e3d26f5cb17b95143ff1037bbea746c1cdaa2f87cb9336dc07f1abd04c4fd150"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.504729 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.516174 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.517899 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.017886173 +0000 UTC m=+151.973247219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.525086 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" event={"ID":"9540e997-8c29-41ba-a743-93c1c62c5fcb","Type":"ContainerStarted","Data":"75ae9e29be74d3e16e1a14b32d2fdd5726f9a6429904161f556405afd8624fb8"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.525235 4733 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-frmrq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.525274 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" podUID="ca36c16c-7eba-43dd-a88c-5f9d18b26d00" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.535314 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" podStartSLOduration=125.535298242 podStartE2EDuration="2m5.535298242s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.534710536 +0000 UTC m=+151.490071602" watchObservedRunningTime="2025-12-04 17:41:29.535298242 +0000 UTC m=+151.490659288" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.537563 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-v9mkr" podStartSLOduration=126.537553942 podStartE2EDuration="2m6.537553942s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.516702182 +0000 UTC m=+151.472063228" watchObservedRunningTime="2025-12-04 17:41:29.537553942 +0000 UTC m=+151.492914988" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.542732 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b93d10d44c680a4726a1bc7d0be7da16a99e4f7824847cc247ece63a45b35627"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.542770 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"99c399baa2a9db8f41f52cc576597720af5fd21c6143c2ea12b20b037e57b807"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.543279 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.558926 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:29 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:29 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:29 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.558970 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.559601 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" event={"ID":"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2","Type":"ContainerStarted","Data":"25dca8e9b9795a403306ba6b4dceaaa7388345943b1590217b9736f1f6151041"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.587172 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" event={"ID":"2cc37bbd-c101-44f0-a327-d6dc3012e375","Type":"ContainerStarted","Data":"78a7e3b8ae56635e49814eec59797b2cfbeeca18caf27d055b271a32ebdf3ff2"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.587215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" event={"ID":"2cc37bbd-c101-44f0-a327-d6dc3012e375","Type":"ContainerStarted","Data":"afb40139a13f3313bd5ed185b430f1db130fae41fc5a238f30c5c425c0de8010"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.611284 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" event={"ID":"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067","Type":"ContainerStarted","Data":"ff4082208165ccdd14f00230bc1024119c8700bacda6ebd367adffa949a13900"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.611325 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" event={"ID":"c99f2792-4e5a-4b3f-9b2b-1b7a8e7e7067","Type":"ContainerStarted","Data":"4fe43a3f596afe6cdb84f37dcd1d1dd1dd1ae5db2e13c32a484c57bc770b63a9"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.616856 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.617736 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.117721768 +0000 UTC m=+152.073082814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.644898 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2gzm8" event={"ID":"35292c5a-4dcc-44d6-8c3b-c1640378a894","Type":"ContainerStarted","Data":"4dbc988a3c1600927473ab6202cf87e6c52548449956f783b64d3dd13beab7e3"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.654762 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" event={"ID":"86e1c39b-aa71-43dc-b331-a66850269dc0","Type":"ContainerStarted","Data":"9e3342aec2b291292aeeb855b686eff8bccbfd9f6591785ed0d90bfa81f37ac7"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.694294 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zzvkk" podStartSLOduration=126.694275636 podStartE2EDuration="2m6.694275636s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.642194056 +0000 UTC m=+151.597555112" watchObservedRunningTime="2025-12-04 17:41:29.694275636 +0000 UTC m=+151.649636682" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.694431 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-khb78" podStartSLOduration=126.694407149 podStartE2EDuration="2m6.694407149s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.692659903 +0000 UTC m=+151.648020949" watchObservedRunningTime="2025-12-04 17:41:29.694407149 +0000 UTC m=+151.649768215" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.732087 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.732743 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.23272799 +0000 UTC m=+152.188089036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.773453 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6adcb5a77dc725b4a20c19c2c5428d3be411467eb0591743255ef118f2d6f622"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.788782 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" event={"ID":"06e612af-44a0-40fe-aeb4-143d2e19c51d","Type":"ContainerStarted","Data":"05df671d043ec449180c2788a6e26af2535d3703fbd12294d2d2dbdf5b4fe36a"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.789134 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" event={"ID":"06e612af-44a0-40fe-aeb4-143d2e19c51d","Type":"ContainerStarted","Data":"3c846b71ec4553cd4409165b31a682bf75a3109fef3ca083beea25c7fd21f9b1"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.796892 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" event={"ID":"c670b992-b3cc-4495-ad15-2b1ec738b3a6","Type":"ContainerStarted","Data":"0fa9f5f523bf260c3ba1a163af8884ea0dc5e501cc98fbfa126bcbe8ca160e64"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.836475 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.836802 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.336773737 +0000 UTC m=+152.292134783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.840546 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jz45r" podStartSLOduration=126.840525738 podStartE2EDuration="2m6.840525738s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.837358853 +0000 UTC m=+151.792719899" watchObservedRunningTime="2025-12-04 17:41:29.840525738 +0000 UTC m=+151.795886784" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.849157 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" event={"ID":"e3966d4d-94a4-4059-a535-fe2eb759c1cf","Type":"ContainerStarted","Data":"8f9341267623846fdc92b4faaadf9e6aae9c94e8aef30e24bc07631faef14d59"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.862539 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" event={"ID":"624f7b69-b973-4294-8043-8747bf3d256f","Type":"ContainerStarted","Data":"d0470e3e39211b31e28ebeb5d8856d26943ae044098cc85879b0ccd8184431c3"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.876968 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" event={"ID":"4a92a269-34fd-4e5a-88b2-9a020e651909","Type":"ContainerStarted","Data":"8b6f474eb7051485eaeaa6dd9342e97abc103f51346ed21fc304fa033ecf683b"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.877036 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" event={"ID":"4a92a269-34fd-4e5a-88b2-9a020e651909","Type":"ContainerStarted","Data":"5cf49528c431885c418be18880df2a961458c6a1058f5b68cbe5d34f5f95d3a8"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.896813 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2b26n" event={"ID":"20a96bf5-566b-4da4-8869-0d27c4b9ef3c","Type":"ContainerStarted","Data":"5fa1de1ffcca8ebe22342b6ee1f25209afbff1166bb393a25f0092d9589548f2"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.909724 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.910173 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.910215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"64e31e4dcf82280751f2f60bd950f3f19d6e6bf195c864ff1775eb9fe9eda0e7"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.910237 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"870e5399294c4f68f1577d3d4e980df9c497fbe34e974f542a6bce9abec6fccd"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.930762 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-nb862" podStartSLOduration=126.930744954 podStartE2EDuration="2m6.930744954s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.929553072 +0000 UTC m=+151.884914118" watchObservedRunningTime="2025-12-04 17:41:29.930744954 +0000 UTC m=+151.886105990" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.931122 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" podStartSLOduration=126.931118374 podStartE2EDuration="2m6.931118374s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.891464358 +0000 UTC m=+151.846825404" watchObservedRunningTime="2025-12-04 17:41:29.931118374 +0000 UTC m=+151.886479420" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.937658 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:29 crc kubenswrapper[4733]: E1204 17:41:29.938118 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.438102472 +0000 UTC m=+152.393463518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.954720 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.958938 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" event={"ID":"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a","Type":"ContainerStarted","Data":"f2cff152a2dbc1a0790690727ac4de2377a389b00a9687ecced7f030553eba19"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.958996 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" event={"ID":"d11858a1-e21b-4dfd-bfa4-5cde50b14b6a","Type":"ContainerStarted","Data":"dc7a3076af73dba1d2183d597725af2b7ac0f3a30e290f151a1f7b35b917cad3"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.970619 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mcr28" podStartSLOduration=126.970596256 podStartE2EDuration="2m6.970596256s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.966616609 +0000 UTC m=+151.921977665" watchObservedRunningTime="2025-12-04 17:41:29.970596256 +0000 UTC m=+151.925957302" Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.972236 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" event={"ID":"1c147058-60bd-403a-b3c6-47750b8e40a6","Type":"ContainerStarted","Data":"c188542267454f35b5955d395bb197efdc5dc1fa6abb54b7339b1a68657cd88a"} Dec 04 17:41:29 crc kubenswrapper[4733]: I1204 17:41:29.983858 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qgps" podStartSLOduration=125.983846192 podStartE2EDuration="2m5.983846192s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:29.982898806 +0000 UTC m=+151.938259852" watchObservedRunningTime="2025-12-04 17:41:29.983846192 +0000 UTC m=+151.939207238" Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.000977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" event={"ID":"e6f87509-6a44-4fbd-85f8-b579f3bff55a","Type":"ContainerStarted","Data":"e0cc79b5953dda6ae9eb3d658248c3c72a2bbdd2b3882417bf200bd70e507d12"} Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.038343 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.039606 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.539587801 +0000 UTC m=+152.494948847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.048499 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" event={"ID":"7767c7f9-42fd-4525-903b-3e6073dd694f","Type":"ContainerStarted","Data":"b3da5eddb481b11a46b227043f3716bdbe43b826358becbbf4c960a2157e0ec7"} Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.071983 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" event={"ID":"70ad32ad-a01a-4676-96f0-399bb8afcbfd","Type":"ContainerStarted","Data":"4cb03591b7bb026af7246c99e342c0fca6ddeb1c0d986d1189b250d0ed8e6c70"} Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.072030 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" event={"ID":"70ad32ad-a01a-4676-96f0-399bb8afcbfd","Type":"ContainerStarted","Data":"ffff1d209c182bc9fe64a1686de77ce7c9292741c4d818c6e000dd8058310b09"} Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.093489 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" event={"ID":"79f2ad87-1b0d-40ea-8bf4-51fb26df839f","Type":"ContainerStarted","Data":"36a1b023ab76f5775401a740bdf593563e4a29498a2b4b871be555b597f0d0de"} Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.106776 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w57sg" Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.129113 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.139378 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.140358 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.64034693 +0000 UTC m=+152.595707976 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.153414 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5qnjl" podStartSLOduration=127.15339961 podStartE2EDuration="2m7.15339961s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:30.152511707 +0000 UTC m=+152.107872773" watchObservedRunningTime="2025-12-04 17:41:30.15339961 +0000 UTC m=+152.108760656" Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.154948 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" podStartSLOduration=126.154940852 podStartE2EDuration="2m6.154940852s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:30.10911743 +0000 UTC m=+152.064478476" watchObservedRunningTime="2025-12-04 17:41:30.154940852 +0000 UTC m=+152.110301898" Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.242204 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.242420 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.742394234 +0000 UTC m=+152.697755280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.242709 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.245173 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.745154848 +0000 UTC m=+152.700515894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.344270 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.344519 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.844504358 +0000 UTC m=+152.799865404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.445703 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.446007 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:30.945996068 +0000 UTC m=+152.901357114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.532949 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:30 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:30 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:30 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.533170 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.546881 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.547200 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.047184339 +0000 UTC m=+153.002545385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.648477 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.648847 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.148829901 +0000 UTC m=+153.104190947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.749953 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.750042 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.250023852 +0000 UTC m=+153.205384898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.750165 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.750443 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.250436303 +0000 UTC m=+153.205797339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.850720 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.850920 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.350894324 +0000 UTC m=+153.306255370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.851174 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.851476 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.351465869 +0000 UTC m=+153.306826915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.952253 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.952443 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.452418193 +0000 UTC m=+153.407779239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:30 crc kubenswrapper[4733]: I1204 17:41:30.952625 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:30 crc kubenswrapper[4733]: E1204 17:41:30.952986 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.452975259 +0000 UTC m=+153.408336305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.053488 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:31 crc kubenswrapper[4733]: E1204 17:41:31.053621 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.553603705 +0000 UTC m=+153.508964841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.053717 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:31 crc kubenswrapper[4733]: E1204 17:41:31.054073 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.554065827 +0000 UTC m=+153.509426873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.108598 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" event={"ID":"1c147058-60bd-403a-b3c6-47750b8e40a6","Type":"ContainerStarted","Data":"f72f3f963c74392e418a978843cc6bd690d71178f7dc1621de11523647ecf8b6"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.109929 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.123195 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" event={"ID":"ae851a3f-c43e-462e-b083-f8d2ba558f55","Type":"ContainerStarted","Data":"9958d699a190321ffee4b80206d755daa1893938714280986a2983b478ad946d"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.124091 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.130111 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" podStartSLOduration=127.130097141 podStartE2EDuration="2m7.130097141s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.128516068 +0000 UTC m=+153.083877124" watchObservedRunningTime="2025-12-04 17:41:31.130097141 +0000 UTC m=+153.085458187" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.130360 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" podStartSLOduration=128.130355868 podStartE2EDuration="2m8.130355868s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:30.241963342 +0000 UTC m=+152.197324398" watchObservedRunningTime="2025-12-04 17:41:31.130355868 +0000 UTC m=+153.085716914" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.132621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" event={"ID":"fa18a76d-2bc5-4e26-956d-fc9d7e5f1cb2","Type":"ContainerStarted","Data":"ce4cef23c6b1625faac04e3f52a2d452f1d0b72e0082b91c0eb85caf3462293a"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.153483 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-872sg" event={"ID":"b55dc242-06d5-43dc-bc4c-b9febdbbd7ae","Type":"ContainerStarted","Data":"dd01e994dbff617e8e489397656159074184cdc5f163835df8a14c74c7f2f33a"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.154415 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:31 crc kubenswrapper[4733]: E1204 17:41:31.154825 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.654791335 +0000 UTC m=+153.610152381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.156150 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" podStartSLOduration=128.156134621 podStartE2EDuration="2m8.156134621s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.153591683 +0000 UTC m=+153.108952739" watchObservedRunningTime="2025-12-04 17:41:31.156134621 +0000 UTC m=+153.111495667" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.160321 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-226s9" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.162407 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-v7zc2" event={"ID":"7767c7f9-42fd-4525-903b-3e6073dd694f","Type":"ContainerStarted","Data":"345f86a845da022b9377bd04cc10b9b769c37f1fcc3628312adba9f8d1fdc2c1"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.171233 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" event={"ID":"4f8163e4-ba70-4cb2-b719-a52e1801ab74","Type":"ContainerStarted","Data":"af45a56b2692b5f7d30991b7d415f7fa756bfe36e6ff1ae691de73dda2368c06"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.186671 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" event={"ID":"aadecb45-b22c-4202-b37b-ba3d071883e9","Type":"ContainerStarted","Data":"2925fcb406a95bdcd818cf2cf1b89fd11aa901b2537027abb42592b39f3c7db4"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.186719 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" event={"ID":"aadecb45-b22c-4202-b37b-ba3d071883e9","Type":"ContainerStarted","Data":"8cb55a2be6f2cf4ed1e8df3edfd4d33152cb429dca6264888684ddaff6a39bc4"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.186930 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.191390 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kpfdb" event={"ID":"86d3ee16-a81a-414f-86a7-77f1ebfa2d40","Type":"ContainerStarted","Data":"cee53088a9a66b70b2e88cf0dd869fd88beb71eb5cde712c61dd1316d935576f"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.197778 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" event={"ID":"c670b992-b3cc-4495-ad15-2b1ec738b3a6","Type":"ContainerStarted","Data":"d3735b01bc40d33c47b2813e8e347d059200ecf8d10255eb078c172312455c33"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.197830 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" event={"ID":"c670b992-b3cc-4495-ad15-2b1ec738b3a6","Type":"ContainerStarted","Data":"689ad7a10756ac9759eb61218c02df8c90b20ed2b54813c740f25e3b22e3be44"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.207088 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" event={"ID":"23c1be98-c468-4a48-9d55-dfaeefef722c","Type":"ContainerStarted","Data":"dadcbbe3eccda09df30cc58f849a4211d3bee1706e1d5eb211bf7ddbafa7d480"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.207516 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.212182 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-5fzhh" podStartSLOduration=128.212169208 podStartE2EDuration="2m8.212169208s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.180782584 +0000 UTC m=+153.136143630" watchObservedRunningTime="2025-12-04 17:41:31.212169208 +0000 UTC m=+153.167530254" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.221625 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" event={"ID":"e6f87509-6a44-4fbd-85f8-b579f3bff55a","Type":"ContainerStarted","Data":"d642307b8bc633d98da8bd3f772c38e8b2ea77f004a7bf9458d8b943f2fac234"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.221673 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" event={"ID":"e6f87509-6a44-4fbd-85f8-b579f3bff55a","Type":"ContainerStarted","Data":"f8d59c01b8d080903066f3872ddc87a48015634944fd283fe2a075854ca2acd2"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.233104 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" event={"ID":"86e1c39b-aa71-43dc-b331-a66850269dc0","Type":"ContainerStarted","Data":"27ff2275e94517f50f0383b11d7679d77fbf90bfd2bfc4afbf8fb8982e3062b0"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.241640 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" podStartSLOduration=128.24162711 podStartE2EDuration="2m8.24162711s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.241288861 +0000 UTC m=+153.196649907" watchObservedRunningTime="2025-12-04 17:41:31.24162711 +0000 UTC m=+153.196988146" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.242038 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-872sg" podStartSLOduration=127.242033431 podStartE2EDuration="2m7.242033431s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.214081629 +0000 UTC m=+153.169442675" watchObservedRunningTime="2025-12-04 17:41:31.242033431 +0000 UTC m=+153.197394477" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.243217 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rzbdr" event={"ID":"b156d126-a527-43ad-a025-cac37696669a","Type":"ContainerStarted","Data":"e73312a551d396d92872f76414c583ab7a242730273716cdc6737ff0f85f9c9a"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.243264 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rzbdr" event={"ID":"b156d126-a527-43ad-a025-cac37696669a","Type":"ContainerStarted","Data":"d3d8023e30a63db7eb1c0e641643d75dec889e3fa0ce96552b5534a654400367"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.243867 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.248182 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" event={"ID":"9540e997-8c29-41ba-a743-93c1c62c5fcb","Type":"ContainerStarted","Data":"692d783fbe4949e746bbe15966da91c00e19929546581bd466947ada0a71a35b"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.248821 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.250551 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" event={"ID":"c5ee72b2-5573-41a9-9c53-8cd27878b57c","Type":"ContainerStarted","Data":"97034d3b16b237b6a5c36de0c739b1fc83848bd083e04c03999a14757677e6e4"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.253105 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" event={"ID":"79f2ad87-1b0d-40ea-8bf4-51fb26df839f","Type":"ContainerStarted","Data":"d24abf46b2c44c997d119b503f9ba47bad76b3c36c274ef3e41aae3fd7aab77c"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.257643 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.258697 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kpfdb" podStartSLOduration=9.258686249 podStartE2EDuration="9.258686249s" podCreationTimestamp="2025-12-04 17:41:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.254745812 +0000 UTC m=+153.210106868" watchObservedRunningTime="2025-12-04 17:41:31.258686249 +0000 UTC m=+153.214047305" Dec 04 17:41:31 crc kubenswrapper[4733]: E1204 17:41:31.259031 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.759018678 +0000 UTC m=+153.714379864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.269673 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a5fedc5df37f99a024d18e3d27d8d49fac5beb07ed771347920f47da7f151d2b"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.288997 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2b26n" event={"ID":"20a96bf5-566b-4da4-8869-0d27c4b9ef3c","Type":"ContainerStarted","Data":"e4efc6071400d423ae0cc31aa089a48eb7f666dc977eb761205234acfa1bd107"} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.289044 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2b26n" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.289140 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vghxl" podStartSLOduration=128.289129997 podStartE2EDuration="2m8.289129997s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.287156944 +0000 UTC m=+153.242517990" watchObservedRunningTime="2025-12-04 17:41:31.289129997 +0000 UTC m=+153.244491043" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.305869 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-frmrq" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.326809 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" podStartSLOduration=127.326763729 podStartE2EDuration="2m7.326763729s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.323387508 +0000 UTC m=+153.278748554" watchObservedRunningTime="2025-12-04 17:41:31.326763729 +0000 UTC m=+153.282124775" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.335928 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b26n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.335983 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2b26n" podUID="20a96bf5-566b-4da4-8869-0d27c4b9ef3c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.359954 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:31 crc kubenswrapper[4733]: E1204 17:41:31.361671 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.861653536 +0000 UTC m=+153.817014592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.383050 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7rff2" podStartSLOduration=128.383037202 podStartE2EDuration="2m8.383037202s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.381106709 +0000 UTC m=+153.336467755" watchObservedRunningTime="2025-12-04 17:41:31.383037202 +0000 UTC m=+153.338398248" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.420781 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5rpdj" podStartSLOduration=128.420763227 podStartE2EDuration="2m8.420763227s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.419229045 +0000 UTC m=+153.374590091" watchObservedRunningTime="2025-12-04 17:41:31.420763227 +0000 UTC m=+153.376124273" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.458583 4733 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.464363 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:31 crc kubenswrapper[4733]: E1204 17:41:31.465700 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 17:41:31.965689465 +0000 UTC m=+153.921050511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rhg5t" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.487912 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rzbdr" podStartSLOduration=9.487891952 podStartE2EDuration="9.487891952s" podCreationTimestamp="2025-12-04 17:41:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.481016106 +0000 UTC m=+153.436377152" watchObservedRunningTime="2025-12-04 17:41:31.487891952 +0000 UTC m=+153.443253018" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.532912 4733 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-04T17:41:31.458607734Z","Handler":null,"Name":""} Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.537047 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:31 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:31 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:31 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.537093 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.542561 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-2b26n" podStartSLOduration=128.54254696 podStartE2EDuration="2m8.54254696s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.541131603 +0000 UTC m=+153.496492649" watchObservedRunningTime="2025-12-04 17:41:31.54254696 +0000 UTC m=+153.497908006" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.574261 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:31 crc kubenswrapper[4733]: E1204 17:41:31.574569 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 17:41:32.074554591 +0000 UTC m=+154.029915637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.576949 4733 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.576978 4733 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.642764 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" podStartSLOduration=127.642746845 podStartE2EDuration="2m7.642746845s" podCreationTimestamp="2025-12-04 17:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.588906937 +0000 UTC m=+153.544267983" watchObservedRunningTime="2025-12-04 17:41:31.642746845 +0000 UTC m=+153.598107891" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.677847 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.707639 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.707676 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.959940 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9c59g" Dec 04 17:41:31 crc kubenswrapper[4733]: I1204 17:41:31.977316 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mmssx" podStartSLOduration=128.977296549 podStartE2EDuration="2m8.977296549s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:31.645755446 +0000 UTC m=+153.601116492" watchObservedRunningTime="2025-12-04 17:41:31.977296549 +0000 UTC m=+153.932657595" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.022032 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rhg5t\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.083002 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.095115 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.124232 4733 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bnzzn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.42:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.124549 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" podUID="ae851a3f-c43e-462e-b083-f8d2ba558f55" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.42:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.132112 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.299056 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" event={"ID":"e6f87509-6a44-4fbd-85f8-b579f3bff55a","Type":"ContainerStarted","Data":"8ca3d3c448b792dad40bda97c3830b622c058a9beec37d0ecaa9cd38e1e6b9e2"} Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.299097 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" event={"ID":"e6f87509-6a44-4fbd-85f8-b579f3bff55a","Type":"ContainerStarted","Data":"710d83c423e3652751bad13c384567871a59efb0759ee331d81b33a90222257b"} Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.306967 4733 generic.go:334] "Generic (PLEG): container finished" podID="79f2ad87-1b0d-40ea-8bf4-51fb26df839f" containerID="d24abf46b2c44c997d119b503f9ba47bad76b3c36c274ef3e41aae3fd7aab77c" exitCode=0 Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.307588 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" event={"ID":"79f2ad87-1b0d-40ea-8bf4-51fb26df839f","Type":"ContainerDied","Data":"d24abf46b2c44c997d119b503f9ba47bad76b3c36c274ef3e41aae3fd7aab77c"} Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.310563 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b26n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.310607 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2b26n" podUID="20a96bf5-566b-4da4-8869-0d27c4b9ef3c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.335496 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-n6k4k" podStartSLOduration=10.33547974 podStartE2EDuration="10.33547974s" podCreationTimestamp="2025-12-04 17:41:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:32.322684286 +0000 UTC m=+154.278045342" watchObservedRunningTime="2025-12-04 17:41:32.33547974 +0000 UTC m=+154.290840786" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.346262 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.354754 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rhg5t"] Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.433400 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.515206 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v2fw4"] Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.516353 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.518091 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.525868 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v2fw4"] Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.535146 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:32 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:32 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:32 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.535193 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.690517 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-catalog-content\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.690698 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-utilities\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.690838 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzdk4\" (UniqueName: \"kubernetes.io/projected/26041ca9-71b9-4deb-a4e0-8192e7bac916-kube-api-access-jzdk4\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.708010 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-29x4q"] Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.708854 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.710925 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.722449 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-29x4q"] Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.792216 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-catalog-content\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.792310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-utilities\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.792394 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzdk4\" (UniqueName: \"kubernetes.io/projected/26041ca9-71b9-4deb-a4e0-8192e7bac916-kube-api-access-jzdk4\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.792866 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-utilities\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.793219 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-catalog-content\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.810633 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzdk4\" (UniqueName: \"kubernetes.io/projected/26041ca9-71b9-4deb-a4e0-8192e7bac916-kube-api-access-jzdk4\") pod \"certified-operators-v2fw4\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.829918 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.893017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-utilities\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.893397 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnlzg\" (UniqueName: \"kubernetes.io/projected/d8eae688-012f-4c73-bd36-828cb9350458-kube-api-access-dnlzg\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.893425 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-catalog-content\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.945376 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cmjzf"] Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.946668 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.951432 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cmjzf"] Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.994691 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-utilities\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.994754 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnlzg\" (UniqueName: \"kubernetes.io/projected/d8eae688-012f-4c73-bd36-828cb9350458-kube-api-access-dnlzg\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.994777 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-catalog-content\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.995438 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-catalog-content\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:32 crc kubenswrapper[4733]: I1204 17:41:32.995671 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-utilities\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.019781 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnlzg\" (UniqueName: \"kubernetes.io/projected/d8eae688-012f-4c73-bd36-828cb9350458-kube-api-access-dnlzg\") pod \"community-operators-29x4q\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.025370 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.055540 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v2fw4"] Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.095627 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-utilities\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.095665 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx4wz\" (UniqueName: \"kubernetes.io/projected/3d1b51bf-b4c8-4a26-9be3-6c06322898db-kube-api-access-bx4wz\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.095694 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-catalog-content\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.108826 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-72rjz"] Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.109730 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.119232 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72rjz"] Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.197085 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-utilities\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.197437 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx4wz\" (UniqueName: \"kubernetes.io/projected/3d1b51bf-b4c8-4a26-9be3-6c06322898db-kube-api-access-bx4wz\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.197479 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-catalog-content\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.197930 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-catalog-content\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.198460 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-utilities\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.214157 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx4wz\" (UniqueName: \"kubernetes.io/projected/3d1b51bf-b4c8-4a26-9be3-6c06322898db-kube-api-access-bx4wz\") pod \"certified-operators-cmjzf\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.261382 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-29x4q"] Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.268810 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.298503 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-catalog-content\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.298554 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-utilities\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.298611 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv46p\" (UniqueName: \"kubernetes.io/projected/7fbdd806-67e0-4af9-ab67-542319ad0d67-kube-api-access-cv46p\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: W1204 17:41:33.305167 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8eae688_012f_4c73_bd36_828cb9350458.slice/crio-ae65f4efa5533f262462cc5107e95b5d5ba7355329c8f27b1e8412232a23264f WatchSource:0}: Error finding container ae65f4efa5533f262462cc5107e95b5d5ba7355329c8f27b1e8412232a23264f: Status 404 returned error can't find the container with id ae65f4efa5533f262462cc5107e95b5d5ba7355329c8f27b1e8412232a23264f Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.315052 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29x4q" event={"ID":"d8eae688-012f-4c73-bd36-828cb9350458","Type":"ContainerStarted","Data":"ae65f4efa5533f262462cc5107e95b5d5ba7355329c8f27b1e8412232a23264f"} Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.316460 4733 generic.go:334] "Generic (PLEG): container finished" podID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerID="6861c904bf894b593b0c32dfe97a422b55ff9699ac08100ecc408df623274cc9" exitCode=0 Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.316870 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2fw4" event={"ID":"26041ca9-71b9-4deb-a4e0-8192e7bac916","Type":"ContainerDied","Data":"6861c904bf894b593b0c32dfe97a422b55ff9699ac08100ecc408df623274cc9"} Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.316894 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2fw4" event={"ID":"26041ca9-71b9-4deb-a4e0-8192e7bac916","Type":"ContainerStarted","Data":"a8888e8306681261b51b11705bd15cf618ef20b2e7f4da8a488642504c2d1427"} Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.321753 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" event={"ID":"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4","Type":"ContainerStarted","Data":"eb49efe3241be5daa80393c404745707c3d93bfd4668941f624d4d394c8e4db5"} Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.321822 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" event={"ID":"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4","Type":"ContainerStarted","Data":"6a566a53da911dc6959bd5233f566fe0c52dd27c85757563ba51c9cedfc5ce3c"} Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.325522 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.352764 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kk5vw" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.363951 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" podStartSLOduration=130.363936972 podStartE2EDuration="2m10.363936972s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:33.359464182 +0000 UTC m=+155.314825228" watchObservedRunningTime="2025-12-04 17:41:33.363936972 +0000 UTC m=+155.319298018" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.400661 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv46p\" (UniqueName: \"kubernetes.io/projected/7fbdd806-67e0-4af9-ab67-542319ad0d67-kube-api-access-cv46p\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.400986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-catalog-content\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.401011 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-utilities\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.401484 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-utilities\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.404417 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-catalog-content\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.422224 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv46p\" (UniqueName: \"kubernetes.io/projected/7fbdd806-67e0-4af9-ab67-542319ad0d67-kube-api-access-cv46p\") pod \"community-operators-72rjz\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.430633 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.505972 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cmjzf"] Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.532112 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:33 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:33 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:33 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.532155 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.572273 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.706770 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkr9k\" (UniqueName: \"kubernetes.io/projected/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-kube-api-access-gkr9k\") pod \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.706842 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-secret-volume\") pod \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.706864 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-config-volume\") pod \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\" (UID: \"79f2ad87-1b0d-40ea-8bf4-51fb26df839f\") " Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.707636 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-config-volume" (OuterVolumeSpecName: "config-volume") pod "79f2ad87-1b0d-40ea-8bf4-51fb26df839f" (UID: "79f2ad87-1b0d-40ea-8bf4-51fb26df839f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.711977 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-kube-api-access-gkr9k" (OuterVolumeSpecName: "kube-api-access-gkr9k") pod "79f2ad87-1b0d-40ea-8bf4-51fb26df839f" (UID: "79f2ad87-1b0d-40ea-8bf4-51fb26df839f"). InnerVolumeSpecName "kube-api-access-gkr9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.712028 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "79f2ad87-1b0d-40ea-8bf4-51fb26df839f" (UID: "79f2ad87-1b0d-40ea-8bf4-51fb26df839f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.808665 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkr9k\" (UniqueName: \"kubernetes.io/projected/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-kube-api-access-gkr9k\") on node \"crc\" DevicePath \"\"" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.809094 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.809116 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79f2ad87-1b0d-40ea-8bf4-51fb26df839f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 17:41:33 crc kubenswrapper[4733]: I1204 17:41:33.874510 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72rjz"] Dec 04 17:41:33 crc kubenswrapper[4733]: W1204 17:41:33.881552 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fbdd806_67e0_4af9_ab67_542319ad0d67.slice/crio-8197ee409ea699159d1899907475918e35627ca6de4afaa753d722e51376d84a WatchSource:0}: Error finding container 8197ee409ea699159d1899907475918e35627ca6de4afaa753d722e51376d84a: Status 404 returned error can't find the container with id 8197ee409ea699159d1899907475918e35627ca6de4afaa753d722e51376d84a Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.331002 4733 generic.go:334] "Generic (PLEG): container finished" podID="d8eae688-012f-4c73-bd36-828cb9350458" containerID="de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e" exitCode=0 Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.331073 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29x4q" event={"ID":"d8eae688-012f-4c73-bd36-828cb9350458","Type":"ContainerDied","Data":"de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e"} Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.334133 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" event={"ID":"79f2ad87-1b0d-40ea-8bf4-51fb26df839f","Type":"ContainerDied","Data":"36a1b023ab76f5775401a740bdf593563e4a29498a2b4b871be555b597f0d0de"} Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.334168 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36a1b023ab76f5775401a740bdf593563e4a29498a2b4b871be555b597f0d0de" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.334188 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.371900 4733 generic.go:334] "Generic (PLEG): container finished" podID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerID="6e661c3146bbd2e2ee8b7946c86d37e94ad934b3a61ad0a8a921d85bb7f8f061" exitCode=0 Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.371951 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmjzf" event={"ID":"3d1b51bf-b4c8-4a26-9be3-6c06322898db","Type":"ContainerDied","Data":"6e661c3146bbd2e2ee8b7946c86d37e94ad934b3a61ad0a8a921d85bb7f8f061"} Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.371999 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmjzf" event={"ID":"3d1b51bf-b4c8-4a26-9be3-6c06322898db","Type":"ContainerStarted","Data":"d80a2dd23723c78826f12cc53561faa7cd7a9a24a50a0710c085329b4f526387"} Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.376024 4733 generic.go:334] "Generic (PLEG): container finished" podID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerID="03b0743770964f297935fbba486ce6c93be169f4022cc2e833aadcbb25e331fb" exitCode=0 Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.376089 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72rjz" event={"ID":"7fbdd806-67e0-4af9-ab67-542319ad0d67","Type":"ContainerDied","Data":"03b0743770964f297935fbba486ce6c93be169f4022cc2e833aadcbb25e331fb"} Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.376117 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72rjz" event={"ID":"7fbdd806-67e0-4af9-ab67-542319ad0d67","Type":"ContainerStarted","Data":"8197ee409ea699159d1899907475918e35627ca6de4afaa753d722e51376d84a"} Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.376700 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.509690 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgmp"] Dec 04 17:41:34 crc kubenswrapper[4733]: E1204 17:41:34.509890 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f2ad87-1b0d-40ea-8bf4-51fb26df839f" containerName="collect-profiles" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.509901 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f2ad87-1b0d-40ea-8bf4-51fb26df839f" containerName="collect-profiles" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.509996 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f2ad87-1b0d-40ea-8bf4-51fb26df839f" containerName="collect-profiles" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.510746 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.516646 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.519546 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgmp"] Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.532468 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:34 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:34 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:34 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.532519 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.596275 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.597501 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.600335 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.600487 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.610716 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.624414 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58xwk\" (UniqueName: \"kubernetes.io/projected/1e13e0b3-c69f-4494-9e11-480d9ced64b8-kube-api-access-58xwk\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.624470 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-utilities\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.624495 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-catalog-content\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.726304 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58xwk\" (UniqueName: \"kubernetes.io/projected/1e13e0b3-c69f-4494-9e11-480d9ced64b8-kube-api-access-58xwk\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.726781 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.726923 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-utilities\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.726952 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-catalog-content\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.727031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.727837 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-catalog-content\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.728550 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-utilities\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.745121 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58xwk\" (UniqueName: \"kubernetes.io/projected/1e13e0b3-c69f-4494-9e11-480d9ced64b8-kube-api-access-58xwk\") pod \"redhat-marketplace-xkgmp\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.804689 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.805066 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.815789 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.833361 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.833459 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.833520 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.868313 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.875092 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.888600 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.888629 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.893220 4733 patch_prober.go:28] interesting pod/console-f9d7485db-rkdfs container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.893271 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rkdfs" podUID="6c99fe8f-586c-4e35-b010-73dad59c59af" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.918639 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.934407 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qxkkc"] Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.939538 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:34 crc kubenswrapper[4733]: I1204 17:41:34.953950 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxkkc"] Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.039338 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-catalog-content\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.039648 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvdrd\" (UniqueName: \"kubernetes.io/projected/1b29e130-61cd-41e5-ab0b-e573e79d96cc-kube-api-access-gvdrd\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.039702 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-utilities\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.141010 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvdrd\" (UniqueName: \"kubernetes.io/projected/1b29e130-61cd-41e5-ab0b-e573e79d96cc-kube-api-access-gvdrd\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.141077 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-utilities\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.141111 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-catalog-content\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.141597 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-catalog-content\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.141665 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-utilities\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.167135 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvdrd\" (UniqueName: \"kubernetes.io/projected/1b29e130-61cd-41e5-ab0b-e573e79d96cc-kube-api-access-gvdrd\") pod \"redhat-marketplace-qxkkc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.189376 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.199229 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgmp"] Dec 04 17:41:35 crc kubenswrapper[4733]: W1204 17:41:35.211762 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e13e0b3_c69f_4494_9e11_480d9ced64b8.slice/crio-300a0b4d47c5054adf89b67e3c7aef638f4c8ce347a62726e7af2fac333fb7e1 WatchSource:0}: Error finding container 300a0b4d47c5054adf89b67e3c7aef638f4c8ce347a62726e7af2fac333fb7e1: Status 404 returned error can't find the container with id 300a0b4d47c5054adf89b67e3c7aef638f4c8ce347a62726e7af2fac333fb7e1 Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.279050 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.388347 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgmp" event={"ID":"1e13e0b3-c69f-4494-9e11-480d9ced64b8","Type":"ContainerStarted","Data":"300a0b4d47c5054adf89b67e3c7aef638f4c8ce347a62726e7af2fac333fb7e1"} Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.391618 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"34e6567e-5012-46a0-9d02-a00f1d70b1d5","Type":"ContainerStarted","Data":"0b8e1c87a0e2154387a1f6a8e3a2c6c5a1d281fcb271a7c923bbaf9c1568e679"} Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.395860 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rs6c6" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.540133 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.559366 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:35 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:35 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:35 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.559417 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.677071 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxkkc"] Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.710486 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gpbpj"] Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.711427 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.713143 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.723151 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gpbpj"] Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.820138 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.846278 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b26n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.846330 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-2b26n container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.846333 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2b26n" podUID="20a96bf5-566b-4da4-8869-0d27c4b9ef3c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.846355 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2b26n" podUID="20a96bf5-566b-4da4-8869-0d27c4b9ef3c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.880077 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-catalog-content\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.880148 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5mqp\" (UniqueName: \"kubernetes.io/projected/c19dfb55-c152-406a-b35e-35498a8b7690-kube-api-access-l5mqp\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:35 crc kubenswrapper[4733]: I1204 17:41:35.880228 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-utilities\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.002301 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5mqp\" (UniqueName: \"kubernetes.io/projected/c19dfb55-c152-406a-b35e-35498a8b7690-kube-api-access-l5mqp\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.002365 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-utilities\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.002425 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-catalog-content\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.003139 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-utilities\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.006089 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-catalog-content\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.031630 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5mqp\" (UniqueName: \"kubernetes.io/projected/c19dfb55-c152-406a-b35e-35498a8b7690-kube-api-access-l5mqp\") pod \"redhat-operators-gpbpj\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.041144 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.108780 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bxf9s"] Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.109717 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.115667 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bxf9s"] Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.204526 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-catalog-content\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.204873 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-utilities\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.204921 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl7t7\" (UniqueName: \"kubernetes.io/projected/2d42ea64-9dc6-4348-a5aa-f72c3f334775-kube-api-access-wl7t7\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.241127 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.241855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.244370 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.245225 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.246159 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.306006 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-catalog-content\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.306100 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-utilities\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.306138 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.306177 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.306199 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl7t7\" (UniqueName: \"kubernetes.io/projected/2d42ea64-9dc6-4348-a5aa-f72c3f334775-kube-api-access-wl7t7\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.306527 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-catalog-content\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.306823 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-utilities\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.322916 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl7t7\" (UniqueName: \"kubernetes.io/projected/2d42ea64-9dc6-4348-a5aa-f72c3f334775-kube-api-access-wl7t7\") pod \"redhat-operators-bxf9s\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.416389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.416452 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.416689 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.425395 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.432505 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.446236 4733 generic.go:334] "Generic (PLEG): container finished" podID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerID="2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb" exitCode=0 Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.446333 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgmp" event={"ID":"1e13e0b3-c69f-4494-9e11-480d9ced64b8","Type":"ContainerDied","Data":"2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb"} Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.450216 4733 generic.go:334] "Generic (PLEG): container finished" podID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerID="f3d2e0fd1c792e22518b390173aaff1a3afeba2d30926e7575dd0390d40a40f4" exitCode=0 Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.450250 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxkkc" event={"ID":"1b29e130-61cd-41e5-ab0b-e573e79d96cc","Type":"ContainerDied","Data":"f3d2e0fd1c792e22518b390173aaff1a3afeba2d30926e7575dd0390d40a40f4"} Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.450302 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxkkc" event={"ID":"1b29e130-61cd-41e5-ab0b-e573e79d96cc","Type":"ContainerStarted","Data":"ee9a90a20acdd0b6bdbd38e088ba0569cf113321dc942ebb2432101e463b217e"} Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.452151 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"34e6567e-5012-46a0-9d02-a00f1d70b1d5","Type":"ContainerStarted","Data":"40f06b182ab98c48e5bde7a9f5e6c056d1fbbb61b1b505caf89266846e13943b"} Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.476886 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.476867869 podStartE2EDuration="2.476867869s" podCreationTimestamp="2025-12-04 17:41:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:36.474095695 +0000 UTC m=+158.429456741" watchObservedRunningTime="2025-12-04 17:41:36.476867869 +0000 UTC m=+158.432228915" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.532458 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:36 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:36 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:36 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.532516 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.548703 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gpbpj"] Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.564874 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:36 crc kubenswrapper[4733]: W1204 17:41:36.636445 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc19dfb55_c152_406a_b35e_35498a8b7690.slice/crio-797f5b525049d270e0df6192f142c6b2e334f7717b043754778c5a59a38f70fb WatchSource:0}: Error finding container 797f5b525049d270e0df6192f142c6b2e334f7717b043754778c5a59a38f70fb: Status 404 returned error can't find the container with id 797f5b525049d270e0df6192f142c6b2e334f7717b043754778c5a59a38f70fb Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.747756 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bxf9s"] Dec 04 17:41:36 crc kubenswrapper[4733]: W1204 17:41:36.783029 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d42ea64_9dc6_4348_a5aa_f72c3f334775.slice/crio-0aa1887f18131d19eca21fd98661cd5ced5db7a99017e2a3cd63055cd35eebb6 WatchSource:0}: Error finding container 0aa1887f18131d19eca21fd98661cd5ced5db7a99017e2a3cd63055cd35eebb6: Status 404 returned error can't find the container with id 0aa1887f18131d19eca21fd98661cd5ced5db7a99017e2a3cd63055cd35eebb6 Dec 04 17:41:36 crc kubenswrapper[4733]: I1204 17:41:36.835870 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 17:41:36 crc kubenswrapper[4733]: W1204 17:41:36.878296 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod6ee5a002_dfe1_4395_b4d3_743a1ee30367.slice/crio-70272bd6b792ace5aca80025381e8c335fd019b98a1a2bd76c9f517271009cdb WatchSource:0}: Error finding container 70272bd6b792ace5aca80025381e8c335fd019b98a1a2bd76c9f517271009cdb: Status 404 returned error can't find the container with id 70272bd6b792ace5aca80025381e8c335fd019b98a1a2bd76c9f517271009cdb Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.470270 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerID="3054e1311a9fbb1200adfaac378f141053d07662b4037199d3d7b0d36080e682" exitCode=0 Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.470538 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bxf9s" event={"ID":"2d42ea64-9dc6-4348-a5aa-f72c3f334775","Type":"ContainerDied","Data":"3054e1311a9fbb1200adfaac378f141053d07662b4037199d3d7b0d36080e682"} Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.470566 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bxf9s" event={"ID":"2d42ea64-9dc6-4348-a5aa-f72c3f334775","Type":"ContainerStarted","Data":"0aa1887f18131d19eca21fd98661cd5ced5db7a99017e2a3cd63055cd35eebb6"} Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.473916 4733 generic.go:334] "Generic (PLEG): container finished" podID="c19dfb55-c152-406a-b35e-35498a8b7690" containerID="0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6" exitCode=0 Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.474005 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpbpj" event={"ID":"c19dfb55-c152-406a-b35e-35498a8b7690","Type":"ContainerDied","Data":"0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6"} Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.474028 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpbpj" event={"ID":"c19dfb55-c152-406a-b35e-35498a8b7690","Type":"ContainerStarted","Data":"797f5b525049d270e0df6192f142c6b2e334f7717b043754778c5a59a38f70fb"} Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.479776 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6ee5a002-dfe1-4395-b4d3-743a1ee30367","Type":"ContainerStarted","Data":"9bd99361573b08370d38fd554d32d7b50960165259514ef2ce8b163c95644a55"} Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.479813 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6ee5a002-dfe1-4395-b4d3-743a1ee30367","Type":"ContainerStarted","Data":"70272bd6b792ace5aca80025381e8c335fd019b98a1a2bd76c9f517271009cdb"} Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.516420 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.51639436 podStartE2EDuration="1.51639436s" podCreationTimestamp="2025-12-04 17:41:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:41:37.516325547 +0000 UTC m=+159.471686593" watchObservedRunningTime="2025-12-04 17:41:37.51639436 +0000 UTC m=+159.471755406" Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.524199 4733 generic.go:334] "Generic (PLEG): container finished" podID="34e6567e-5012-46a0-9d02-a00f1d70b1d5" containerID="40f06b182ab98c48e5bde7a9f5e6c056d1fbbb61b1b505caf89266846e13943b" exitCode=0 Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.524933 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"34e6567e-5012-46a0-9d02-a00f1d70b1d5","Type":"ContainerDied","Data":"40f06b182ab98c48e5bde7a9f5e6c056d1fbbb61b1b505caf89266846e13943b"} Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.532260 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:37 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:37 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:37 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:37 crc kubenswrapper[4733]: I1204 17:41:37.532306 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:38 crc kubenswrapper[4733]: I1204 17:41:38.532898 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:38 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:38 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:38 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:38 crc kubenswrapper[4733]: I1204 17:41:38.532946 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:38 crc kubenswrapper[4733]: I1204 17:41:38.534624 4733 generic.go:334] "Generic (PLEG): container finished" podID="6ee5a002-dfe1-4395-b4d3-743a1ee30367" containerID="9bd99361573b08370d38fd554d32d7b50960165259514ef2ce8b163c95644a55" exitCode=0 Dec 04 17:41:38 crc kubenswrapper[4733]: I1204 17:41:38.534770 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6ee5a002-dfe1-4395-b4d3-743a1ee30367","Type":"ContainerDied","Data":"9bd99361573b08370d38fd554d32d7b50960165259514ef2ce8b163c95644a55"} Dec 04 17:41:39 crc kubenswrapper[4733]: I1204 17:41:39.531256 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:39 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:39 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:39 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:39 crc kubenswrapper[4733]: I1204 17:41:39.531310 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:40 crc kubenswrapper[4733]: I1204 17:41:40.531875 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:40 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:40 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:40 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:40 crc kubenswrapper[4733]: I1204 17:41:40.532299 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:40 crc kubenswrapper[4733]: I1204 17:41:40.802556 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rzbdr" Dec 04 17:41:41 crc kubenswrapper[4733]: I1204 17:41:41.531085 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:41 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:41 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:41 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:41 crc kubenswrapper[4733]: I1204 17:41:41.531140 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:42 crc kubenswrapper[4733]: I1204 17:41:42.531940 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:42 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:42 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:42 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:42 crc kubenswrapper[4733]: I1204 17:41:42.532007 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:43 crc kubenswrapper[4733]: I1204 17:41:43.531076 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:43 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:43 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:43 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:43 crc kubenswrapper[4733]: I1204 17:41:43.531153 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.533716 4733 patch_prober.go:28] interesting pod/router-default-5444994796-87dsb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 17:41:44 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Dec 04 17:41:44 crc kubenswrapper[4733]: [+]process-running ok Dec 04 17:41:44 crc kubenswrapper[4733]: healthz check failed Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.534066 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-87dsb" podUID="8e123c23-4206-4c39-8549-9cf6727aff8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.548719 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.558391 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.593068 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6ee5a002-dfe1-4395-b4d3-743a1ee30367","Type":"ContainerDied","Data":"70272bd6b792ace5aca80025381e8c335fd019b98a1a2bd76c9f517271009cdb"} Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.593119 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70272bd6b792ace5aca80025381e8c335fd019b98a1a2bd76c9f517271009cdb" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.593118 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.598659 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"34e6567e-5012-46a0-9d02-a00f1d70b1d5","Type":"ContainerDied","Data":"0b8e1c87a0e2154387a1f6a8e3a2c6c5a1d281fcb271a7c923bbaf9c1568e679"} Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.598704 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b8e1c87a0e2154387a1f6a8e3a2c6c5a1d281fcb271a7c923bbaf9c1568e679" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.598839 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.644232 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kube-api-access\") pod \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\" (UID: \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\") " Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.644343 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kubelet-dir\") pod \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\" (UID: \"34e6567e-5012-46a0-9d02-a00f1d70b1d5\") " Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.644372 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kubelet-dir\") pod \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\" (UID: \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\") " Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.644401 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kube-api-access\") pod \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\" (UID: \"6ee5a002-dfe1-4395-b4d3-743a1ee30367\") " Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.644467 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "34e6567e-5012-46a0-9d02-a00f1d70b1d5" (UID: "34e6567e-5012-46a0-9d02-a00f1d70b1d5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.644491 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6ee5a002-dfe1-4395-b4d3-743a1ee30367" (UID: "6ee5a002-dfe1-4395-b4d3-743a1ee30367"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.645414 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.645441 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.650581 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6ee5a002-dfe1-4395-b4d3-743a1ee30367" (UID: "6ee5a002-dfe1-4395-b4d3-743a1ee30367"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.655066 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "34e6567e-5012-46a0-9d02-a00f1d70b1d5" (UID: "34e6567e-5012-46a0-9d02-a00f1d70b1d5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.747040 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34e6567e-5012-46a0-9d02-a00f1d70b1d5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.747094 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ee5a002-dfe1-4395-b4d3-743a1ee30367-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.889214 4733 patch_prober.go:28] interesting pod/console-f9d7485db-rkdfs container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 04 17:41:44 crc kubenswrapper[4733]: I1204 17:41:44.889322 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rkdfs" podUID="6c99fe8f-586c-4e35-b010-73dad59c59af" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 04 17:41:45 crc kubenswrapper[4733]: I1204 17:41:45.361848 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:41:45 crc kubenswrapper[4733]: I1204 17:41:45.361912 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:41:45 crc kubenswrapper[4733]: I1204 17:41:45.534534 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:45 crc kubenswrapper[4733]: I1204 17:41:45.539114 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-87dsb" Dec 04 17:41:45 crc kubenswrapper[4733]: I1204 17:41:45.856616 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-2b26n" Dec 04 17:41:46 crc kubenswrapper[4733]: I1204 17:41:46.368026 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:46 crc kubenswrapper[4733]: I1204 17:41:46.373826 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79d2d07f-2f47-4970-b6e5-7d25b02e639e-metrics-certs\") pod \"network-metrics-daemon-mlwj8\" (UID: \"79d2d07f-2f47-4970-b6e5-7d25b02e639e\") " pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:46 crc kubenswrapper[4733]: I1204 17:41:46.474067 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mlwj8" Dec 04 17:41:52 crc kubenswrapper[4733]: I1204 17:41:52.137636 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:41:54 crc kubenswrapper[4733]: I1204 17:41:54.895600 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:41:54 crc kubenswrapper[4733]: I1204 17:41:54.902200 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:42:05 crc kubenswrapper[4733]: I1204 17:42:05.707090 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 17:42:05 crc kubenswrapper[4733]: I1204 17:42:05.721352 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-csj6s" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.055135 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.055293 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jzdk4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-v2fw4_openshift-marketplace(26041ca9-71b9-4deb-a4e0-8192e7bac916): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.056486 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-v2fw4" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.200129 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.200551 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gvdrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-qxkkc_openshift-marketplace(1b29e130-61cd-41e5-ab0b-e573e79d96cc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.201732 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-qxkkc" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.205128 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.205493 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bx4wz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cmjzf_openshift-marketplace(3d1b51bf-b4c8-4a26-9be3-6c06322898db): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 17:42:07 crc kubenswrapper[4733]: E1204 17:42:07.206609 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cmjzf" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" Dec 04 17:42:07 crc kubenswrapper[4733]: I1204 17:42:07.490126 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mlwj8"] Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.047068 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-v2fw4" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.047198 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cmjzf" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" Dec 04 17:42:10 crc kubenswrapper[4733]: W1204 17:42:10.050600 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79d2d07f_2f47_4970_b6e5_7d25b02e639e.slice/crio-96d167a689557deec3380fddb2e82f91c202e360413871714449c210f8217ec0 WatchSource:0}: Error finding container 96d167a689557deec3380fddb2e82f91c202e360413871714449c210f8217ec0: Status 404 returned error can't find the container with id 96d167a689557deec3380fddb2e82f91c202e360413871714449c210f8217ec0 Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.053358 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-qxkkc" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.094253 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.094391 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l5mqp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gpbpj_openshift-marketplace(c19dfb55-c152-406a-b35e-35498a8b7690): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.095772 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gpbpj" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.099387 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.099487 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wl7t7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-bxf9s_openshift-marketplace(2d42ea64-9dc6-4348-a5aa-f72c3f334775): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.100652 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-bxf9s" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.776052 4733 generic.go:334] "Generic (PLEG): container finished" podID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerID="bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad" exitCode=0 Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.776315 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgmp" event={"ID":"1e13e0b3-c69f-4494-9e11-480d9ced64b8","Type":"ContainerDied","Data":"bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad"} Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.782757 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" event={"ID":"79d2d07f-2f47-4970-b6e5-7d25b02e639e","Type":"ContainerStarted","Data":"edca57bd4c256821537d721ea87255222ed6867d5b6e98e91050dd9a03bc6176"} Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.782811 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" event={"ID":"79d2d07f-2f47-4970-b6e5-7d25b02e639e","Type":"ContainerStarted","Data":"4a4ebe6fd112158d8181359c124f0fdf3c8933039e46b97d9e6e764d877dcdf4"} Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.782821 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mlwj8" event={"ID":"79d2d07f-2f47-4970-b6e5-7d25b02e639e","Type":"ContainerStarted","Data":"96d167a689557deec3380fddb2e82f91c202e360413871714449c210f8217ec0"} Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.787666 4733 generic.go:334] "Generic (PLEG): container finished" podID="d8eae688-012f-4c73-bd36-828cb9350458" containerID="45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938" exitCode=0 Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.787722 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29x4q" event={"ID":"d8eae688-012f-4c73-bd36-828cb9350458","Type":"ContainerDied","Data":"45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938"} Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.807714 4733 generic.go:334] "Generic (PLEG): container finished" podID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerID="58be3f40f1748659e4736be30bad5da226a99d495f10aacedee9d723e3618c9a" exitCode=0 Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.808259 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72rjz" event={"ID":"7fbdd806-67e0-4af9-ab67-542319ad0d67","Type":"ContainerDied","Data":"58be3f40f1748659e4736be30bad5da226a99d495f10aacedee9d723e3618c9a"} Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.812124 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gpbpj" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" Dec 04 17:42:10 crc kubenswrapper[4733]: E1204 17:42:10.812461 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-bxf9s" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" Dec 04 17:42:10 crc kubenswrapper[4733]: I1204 17:42:10.826758 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-mlwj8" podStartSLOduration=167.826733331 podStartE2EDuration="2m47.826733331s" podCreationTimestamp="2025-12-04 17:39:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:42:10.825927719 +0000 UTC m=+192.781288775" watchObservedRunningTime="2025-12-04 17:42:10.826733331 +0000 UTC m=+192.782094417" Dec 04 17:42:11 crc kubenswrapper[4733]: I1204 17:42:11.815905 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72rjz" event={"ID":"7fbdd806-67e0-4af9-ab67-542319ad0d67","Type":"ContainerStarted","Data":"705e3611f446f49c906edab79bf2c282ba00e585ecd9bb4799ca4362087d00a8"} Dec 04 17:42:11 crc kubenswrapper[4733]: I1204 17:42:11.819238 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgmp" event={"ID":"1e13e0b3-c69f-4494-9e11-480d9ced64b8","Type":"ContainerStarted","Data":"78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d"} Dec 04 17:42:11 crc kubenswrapper[4733]: I1204 17:42:11.832199 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-72rjz" podStartSLOduration=1.919761423 podStartE2EDuration="38.832181834s" podCreationTimestamp="2025-12-04 17:41:33 +0000 UTC" firstStartedPulling="2025-12-04 17:41:34.377768971 +0000 UTC m=+156.333130017" lastFinishedPulling="2025-12-04 17:42:11.290189382 +0000 UTC m=+193.245550428" observedRunningTime="2025-12-04 17:42:11.830673084 +0000 UTC m=+193.786034150" watchObservedRunningTime="2025-12-04 17:42:11.832181834 +0000 UTC m=+193.787542880" Dec 04 17:42:11 crc kubenswrapper[4733]: I1204 17:42:11.833602 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29x4q" event={"ID":"d8eae688-012f-4c73-bd36-828cb9350458","Type":"ContainerStarted","Data":"aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd"} Dec 04 17:42:11 crc kubenswrapper[4733]: I1204 17:42:11.848760 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xkgmp" podStartSLOduration=3.112623133 podStartE2EDuration="37.84874391s" podCreationTimestamp="2025-12-04 17:41:34 +0000 UTC" firstStartedPulling="2025-12-04 17:41:36.462566205 +0000 UTC m=+158.417927251" lastFinishedPulling="2025-12-04 17:42:11.198686982 +0000 UTC m=+193.154048028" observedRunningTime="2025-12-04 17:42:11.845333208 +0000 UTC m=+193.800694264" watchObservedRunningTime="2025-12-04 17:42:11.84874391 +0000 UTC m=+193.804104956" Dec 04 17:42:11 crc kubenswrapper[4733]: I1204 17:42:11.863032 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-29x4q" podStartSLOduration=2.900271939 podStartE2EDuration="39.862991333s" podCreationTimestamp="2025-12-04 17:41:32 +0000 UTC" firstStartedPulling="2025-12-04 17:41:34.332919635 +0000 UTC m=+156.288280681" lastFinishedPulling="2025-12-04 17:42:11.295639029 +0000 UTC m=+193.251000075" observedRunningTime="2025-12-04 17:42:11.860422124 +0000 UTC m=+193.815783170" watchObservedRunningTime="2025-12-04 17:42:11.862991333 +0000 UTC m=+193.818352429" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.036815 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 17:42:12 crc kubenswrapper[4733]: E1204 17:42:12.037008 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e6567e-5012-46a0-9d02-a00f1d70b1d5" containerName="pruner" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.037019 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e6567e-5012-46a0-9d02-a00f1d70b1d5" containerName="pruner" Dec 04 17:42:12 crc kubenswrapper[4733]: E1204 17:42:12.037036 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee5a002-dfe1-4395-b4d3-743a1ee30367" containerName="pruner" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.037042 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee5a002-dfe1-4395-b4d3-743a1ee30367" containerName="pruner" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.037138 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="34e6567e-5012-46a0-9d02-a00f1d70b1d5" containerName="pruner" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.037150 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ee5a002-dfe1-4395-b4d3-743a1ee30367" containerName="pruner" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.037521 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.040469 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.040566 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.055977 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.149707 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.150042 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.250680 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.250906 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.250773 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.274245 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.360082 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.762058 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 17:42:12 crc kubenswrapper[4733]: W1204 17:42:12.771753 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod365ab5ed_b33a_439d_82f9_4b1d17984fb3.slice/crio-7d482dc39ab4b8d873aa23b35dee184d59800fe15a59b771e3312147f5dd3a72 WatchSource:0}: Error finding container 7d482dc39ab4b8d873aa23b35dee184d59800fe15a59b771e3312147f5dd3a72: Status 404 returned error can't find the container with id 7d482dc39ab4b8d873aa23b35dee184d59800fe15a59b771e3312147f5dd3a72 Dec 04 17:42:12 crc kubenswrapper[4733]: I1204 17:42:12.841306 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"365ab5ed-b33a-439d-82f9-4b1d17984fb3","Type":"ContainerStarted","Data":"7d482dc39ab4b8d873aa23b35dee184d59800fe15a59b771e3312147f5dd3a72"} Dec 04 17:42:13 crc kubenswrapper[4733]: I1204 17:42:13.026634 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:42:13 crc kubenswrapper[4733]: I1204 17:42:13.026665 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:42:13 crc kubenswrapper[4733]: I1204 17:42:13.431754 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:42:13 crc kubenswrapper[4733]: I1204 17:42:13.431957 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:42:13 crc kubenswrapper[4733]: I1204 17:42:13.479297 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:42:13 crc kubenswrapper[4733]: I1204 17:42:13.847680 4733 generic.go:334] "Generic (PLEG): container finished" podID="365ab5ed-b33a-439d-82f9-4b1d17984fb3" containerID="5077fb99e1c19bea6463bd1ce0e797a0dda3b7808035421584b547c0af709034" exitCode=0 Dec 04 17:42:13 crc kubenswrapper[4733]: I1204 17:42:13.847736 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"365ab5ed-b33a-439d-82f9-4b1d17984fb3","Type":"ContainerDied","Data":"5077fb99e1c19bea6463bd1ce0e797a0dda3b7808035421584b547c0af709034"} Dec 04 17:42:14 crc kubenswrapper[4733]: I1204 17:42:14.091460 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-29x4q" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="registry-server" probeResult="failure" output=< Dec 04 17:42:14 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 17:42:14 crc kubenswrapper[4733]: > Dec 04 17:42:14 crc kubenswrapper[4733]: I1204 17:42:14.876357 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:42:14 crc kubenswrapper[4733]: I1204 17:42:14.876397 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:42:14 crc kubenswrapper[4733]: I1204 17:42:14.920330 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.179240 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.287378 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kube-api-access\") pod \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\" (UID: \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\") " Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.287448 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kubelet-dir\") pod \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\" (UID: \"365ab5ed-b33a-439d-82f9-4b1d17984fb3\") " Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.287617 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "365ab5ed-b33a-439d-82f9-4b1d17984fb3" (UID: "365ab5ed-b33a-439d-82f9-4b1d17984fb3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.294953 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "365ab5ed-b33a-439d-82f9-4b1d17984fb3" (UID: "365ab5ed-b33a-439d-82f9-4b1d17984fb3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.361906 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.361965 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.389116 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.389149 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/365ab5ed-b33a-439d-82f9-4b1d17984fb3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.861047 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"365ab5ed-b33a-439d-82f9-4b1d17984fb3","Type":"ContainerDied","Data":"7d482dc39ab4b8d873aa23b35dee184d59800fe15a59b771e3312147f5dd3a72"} Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.861393 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d482dc39ab4b8d873aa23b35dee184d59800fe15a59b771e3312147f5dd3a72" Dec 04 17:42:15 crc kubenswrapper[4733]: I1204 17:42:15.861073 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 17:42:17 crc kubenswrapper[4733]: I1204 17:42:17.832410 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 17:42:17 crc kubenswrapper[4733]: E1204 17:42:17.832688 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="365ab5ed-b33a-439d-82f9-4b1d17984fb3" containerName="pruner" Dec 04 17:42:17 crc kubenswrapper[4733]: I1204 17:42:17.832703 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="365ab5ed-b33a-439d-82f9-4b1d17984fb3" containerName="pruner" Dec 04 17:42:17 crc kubenswrapper[4733]: I1204 17:42:17.832856 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="365ab5ed-b33a-439d-82f9-4b1d17984fb3" containerName="pruner" Dec 04 17:42:17 crc kubenswrapper[4733]: I1204 17:42:17.833308 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:17 crc kubenswrapper[4733]: I1204 17:42:17.835090 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 17:42:17 crc kubenswrapper[4733]: I1204 17:42:17.838651 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 17:42:17 crc kubenswrapper[4733]: I1204 17:42:17.841834 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.022337 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dff4fe6a-a0a3-4832-800b-aa23959389bd-kube-api-access\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.022411 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.022434 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-var-lock\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.123862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.123911 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-var-lock\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.123949 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dff4fe6a-a0a3-4832-800b-aa23959389bd-kube-api-access\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.123977 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.124080 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-var-lock\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.162241 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dff4fe6a-a0a3-4832-800b-aa23959389bd-kube-api-access\") pod \"installer-9-crc\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.460123 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:18 crc kubenswrapper[4733]: I1204 17:42:18.901986 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 17:42:18 crc kubenswrapper[4733]: W1204 17:42:18.913136 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poddff4fe6a_a0a3_4832_800b_aa23959389bd.slice/crio-6beb12c88f1b0e8bd0eea502dc1aaf57c69c5f4da7ee27c494a2c6b8e2b88c44 WatchSource:0}: Error finding container 6beb12c88f1b0e8bd0eea502dc1aaf57c69c5f4da7ee27c494a2c6b8e2b88c44: Status 404 returned error can't find the container with id 6beb12c88f1b0e8bd0eea502dc1aaf57c69c5f4da7ee27c494a2c6b8e2b88c44 Dec 04 17:42:19 crc kubenswrapper[4733]: I1204 17:42:19.888379 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dff4fe6a-a0a3-4832-800b-aa23959389bd","Type":"ContainerStarted","Data":"00c22ee848a6cc76bbd2ada8ccf6328f9493636f7be780ed8d3291c692f7ff0c"} Dec 04 17:42:19 crc kubenswrapper[4733]: I1204 17:42:19.888743 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dff4fe6a-a0a3-4832-800b-aa23959389bd","Type":"ContainerStarted","Data":"6beb12c88f1b0e8bd0eea502dc1aaf57c69c5f4da7ee27c494a2c6b8e2b88c44"} Dec 04 17:42:22 crc kubenswrapper[4733]: I1204 17:42:22.354849 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=5.354835213 podStartE2EDuration="5.354835213s" podCreationTimestamp="2025-12-04 17:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:42:19.903623992 +0000 UTC m=+201.858985048" watchObservedRunningTime="2025-12-04 17:42:22.354835213 +0000 UTC m=+204.310196259" Dec 04 17:42:23 crc kubenswrapper[4733]: I1204 17:42:23.074582 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:42:23 crc kubenswrapper[4733]: I1204 17:42:23.113397 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:42:23 crc kubenswrapper[4733]: I1204 17:42:23.477132 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:42:24 crc kubenswrapper[4733]: I1204 17:42:24.713600 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bnzzn"] Dec 04 17:42:24 crc kubenswrapper[4733]: I1204 17:42:24.925113 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:42:26 crc kubenswrapper[4733]: I1204 17:42:26.933282 4733 generic.go:334] "Generic (PLEG): container finished" podID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerID="026c36ed3b3eedfda502d93a5301bc0142cca40e78c8897ef45d2e2443259a61" exitCode=0 Dec 04 17:42:26 crc kubenswrapper[4733]: I1204 17:42:26.933337 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxkkc" event={"ID":"1b29e130-61cd-41e5-ab0b-e573e79d96cc","Type":"ContainerDied","Data":"026c36ed3b3eedfda502d93a5301bc0142cca40e78c8897ef45d2e2443259a61"} Dec 04 17:42:26 crc kubenswrapper[4733]: I1204 17:42:26.936576 4733 generic.go:334] "Generic (PLEG): container finished" podID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerID="7e5b33371b5190df651ec97ac9d4cc20ac72a8067931ad0c515c343240748cc0" exitCode=0 Dec 04 17:42:26 crc kubenswrapper[4733]: I1204 17:42:26.936642 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2fw4" event={"ID":"26041ca9-71b9-4deb-a4e0-8192e7bac916","Type":"ContainerDied","Data":"7e5b33371b5190df651ec97ac9d4cc20ac72a8067931ad0c515c343240748cc0"} Dec 04 17:42:26 crc kubenswrapper[4733]: I1204 17:42:26.940241 4733 generic.go:334] "Generic (PLEG): container finished" podID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerID="9203b7f314f8ee7f803e63845c5dbc16979ec231a0d83fb4c57d4dde5b90733a" exitCode=0 Dec 04 17:42:26 crc kubenswrapper[4733]: I1204 17:42:26.940301 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmjzf" event={"ID":"3d1b51bf-b4c8-4a26-9be3-6c06322898db","Type":"ContainerDied","Data":"9203b7f314f8ee7f803e63845c5dbc16979ec231a0d83fb4c57d4dde5b90733a"} Dec 04 17:42:26 crc kubenswrapper[4733]: I1204 17:42:26.942716 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bxf9s" event={"ID":"2d42ea64-9dc6-4348-a5aa-f72c3f334775","Type":"ContainerStarted","Data":"610dba7a10b68f0f42dfefda016d20a12e9f91267d19568b6c19f924978fe4c6"} Dec 04 17:42:26 crc kubenswrapper[4733]: I1204 17:42:26.946204 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpbpj" event={"ID":"c19dfb55-c152-406a-b35e-35498a8b7690","Type":"ContainerStarted","Data":"8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44"} Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.564328 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72rjz"] Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.564542 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-72rjz" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerName="registry-server" containerID="cri-o://705e3611f446f49c906edab79bf2c282ba00e585ecd9bb4799ca4362087d00a8" gracePeriod=2 Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.961730 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmjzf" event={"ID":"3d1b51bf-b4c8-4a26-9be3-6c06322898db","Type":"ContainerStarted","Data":"fbf7040307cd01072b5ab00eb0277eea0afac00461d5d9f9d4df8776050d576e"} Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.969607 4733 generic.go:334] "Generic (PLEG): container finished" podID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerID="705e3611f446f49c906edab79bf2c282ba00e585ecd9bb4799ca4362087d00a8" exitCode=0 Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.969664 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72rjz" event={"ID":"7fbdd806-67e0-4af9-ab67-542319ad0d67","Type":"ContainerDied","Data":"705e3611f446f49c906edab79bf2c282ba00e585ecd9bb4799ca4362087d00a8"} Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.971399 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerID="610dba7a10b68f0f42dfefda016d20a12e9f91267d19568b6c19f924978fe4c6" exitCode=0 Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.971592 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bxf9s" event={"ID":"2d42ea64-9dc6-4348-a5aa-f72c3f334775","Type":"ContainerDied","Data":"610dba7a10b68f0f42dfefda016d20a12e9f91267d19568b6c19f924978fe4c6"} Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.975165 4733 generic.go:334] "Generic (PLEG): container finished" podID="c19dfb55-c152-406a-b35e-35498a8b7690" containerID="8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44" exitCode=0 Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.975253 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpbpj" event={"ID":"c19dfb55-c152-406a-b35e-35498a8b7690","Type":"ContainerDied","Data":"8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44"} Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.980282 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cmjzf" podStartSLOduration=2.931868354 podStartE2EDuration="55.980268172s" podCreationTimestamp="2025-12-04 17:41:32 +0000 UTC" firstStartedPulling="2025-12-04 17:41:34.374871693 +0000 UTC m=+156.330232739" lastFinishedPulling="2025-12-04 17:42:27.423271521 +0000 UTC m=+209.378632557" observedRunningTime="2025-12-04 17:42:27.97943619 +0000 UTC m=+209.934797246" watchObservedRunningTime="2025-12-04 17:42:27.980268172 +0000 UTC m=+209.935629218" Dec 04 17:42:27 crc kubenswrapper[4733]: I1204 17:42:27.991548 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxkkc" event={"ID":"1b29e130-61cd-41e5-ab0b-e573e79d96cc","Type":"ContainerStarted","Data":"451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979"} Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.004556 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2fw4" event={"ID":"26041ca9-71b9-4deb-a4e0-8192e7bac916","Type":"ContainerStarted","Data":"8d131100c38d0de1ba07ae88f82c29e47f56606933dbf747d75ad0158efe3f11"} Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.038724 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v2fw4" podStartSLOduration=2.036703903 podStartE2EDuration="56.038710013s" podCreationTimestamp="2025-12-04 17:41:32 +0000 UTC" firstStartedPulling="2025-12-04 17:41:33.325230311 +0000 UTC m=+155.280591367" lastFinishedPulling="2025-12-04 17:42:27.327236421 +0000 UTC m=+209.282597477" observedRunningTime="2025-12-04 17:42:28.035623578 +0000 UTC m=+209.990984634" watchObservedRunningTime="2025-12-04 17:42:28.038710013 +0000 UTC m=+209.994071059" Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.055511 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qxkkc" podStartSLOduration=3.18147253 podStartE2EDuration="54.055486332s" podCreationTimestamp="2025-12-04 17:41:34 +0000 UTC" firstStartedPulling="2025-12-04 17:41:36.464004734 +0000 UTC m=+158.419365780" lastFinishedPulling="2025-12-04 17:42:27.338018516 +0000 UTC m=+209.293379582" observedRunningTime="2025-12-04 17:42:28.052573702 +0000 UTC m=+210.007934748" watchObservedRunningTime="2025-12-04 17:42:28.055486332 +0000 UTC m=+210.010847378" Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.118721 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.251764 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv46p\" (UniqueName: \"kubernetes.io/projected/7fbdd806-67e0-4af9-ab67-542319ad0d67-kube-api-access-cv46p\") pod \"7fbdd806-67e0-4af9-ab67-542319ad0d67\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.252160 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-utilities\") pod \"7fbdd806-67e0-4af9-ab67-542319ad0d67\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.252203 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-catalog-content\") pod \"7fbdd806-67e0-4af9-ab67-542319ad0d67\" (UID: \"7fbdd806-67e0-4af9-ab67-542319ad0d67\") " Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.254588 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-utilities" (OuterVolumeSpecName: "utilities") pod "7fbdd806-67e0-4af9-ab67-542319ad0d67" (UID: "7fbdd806-67e0-4af9-ab67-542319ad0d67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.257155 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbdd806-67e0-4af9-ab67-542319ad0d67-kube-api-access-cv46p" (OuterVolumeSpecName: "kube-api-access-cv46p") pod "7fbdd806-67e0-4af9-ab67-542319ad0d67" (UID: "7fbdd806-67e0-4af9-ab67-542319ad0d67"). InnerVolumeSpecName "kube-api-access-cv46p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.304718 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fbdd806-67e0-4af9-ab67-542319ad0d67" (UID: "7fbdd806-67e0-4af9-ab67-542319ad0d67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.352938 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.352968 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv46p\" (UniqueName: \"kubernetes.io/projected/7fbdd806-67e0-4af9-ab67-542319ad0d67-kube-api-access-cv46p\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:28 crc kubenswrapper[4733]: I1204 17:42:28.352979 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbdd806-67e0-4af9-ab67-542319ad0d67-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.011605 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpbpj" event={"ID":"c19dfb55-c152-406a-b35e-35498a8b7690","Type":"ContainerStarted","Data":"2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8"} Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.013322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bxf9s" event={"ID":"2d42ea64-9dc6-4348-a5aa-f72c3f334775","Type":"ContainerStarted","Data":"cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0"} Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.015709 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72rjz" event={"ID":"7fbdd806-67e0-4af9-ab67-542319ad0d67","Type":"ContainerDied","Data":"8197ee409ea699159d1899907475918e35627ca6de4afaa753d722e51376d84a"} Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.015879 4733 scope.go:117] "RemoveContainer" containerID="705e3611f446f49c906edab79bf2c282ba00e585ecd9bb4799ca4362087d00a8" Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.015763 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72rjz" Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.027928 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gpbpj" podStartSLOduration=3.034047525 podStartE2EDuration="54.02790952s" podCreationTimestamp="2025-12-04 17:41:35 +0000 UTC" firstStartedPulling="2025-12-04 17:41:37.474979936 +0000 UTC m=+159.430340982" lastFinishedPulling="2025-12-04 17:42:28.468841911 +0000 UTC m=+210.424202977" observedRunningTime="2025-12-04 17:42:29.026720707 +0000 UTC m=+210.982081983" watchObservedRunningTime="2025-12-04 17:42:29.02790952 +0000 UTC m=+210.983270566" Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.028426 4733 scope.go:117] "RemoveContainer" containerID="58be3f40f1748659e4736be30bad5da226a99d495f10aacedee9d723e3618c9a" Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.049342 4733 scope.go:117] "RemoveContainer" containerID="03b0743770964f297935fbba486ce6c93be169f4022cc2e833aadcbb25e331fb" Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.056512 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bxf9s" podStartSLOduration=1.963009352 podStartE2EDuration="53.056495183s" podCreationTimestamp="2025-12-04 17:41:36 +0000 UTC" firstStartedPulling="2025-12-04 17:41:37.47175941 +0000 UTC m=+159.427120456" lastFinishedPulling="2025-12-04 17:42:28.565245241 +0000 UTC m=+210.520606287" observedRunningTime="2025-12-04 17:42:29.054764055 +0000 UTC m=+211.010125121" watchObservedRunningTime="2025-12-04 17:42:29.056495183 +0000 UTC m=+211.011856229" Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.071477 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72rjz"] Dec 04 17:42:29 crc kubenswrapper[4733]: I1204 17:42:29.076071 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-72rjz"] Dec 04 17:42:30 crc kubenswrapper[4733]: I1204 17:42:30.345880 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" path="/var/lib/kubelet/pods/7fbdd806-67e0-4af9-ab67-542319ad0d67/volumes" Dec 04 17:42:32 crc kubenswrapper[4733]: I1204 17:42:32.830312 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:42:32 crc kubenswrapper[4733]: I1204 17:42:32.830625 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:42:32 crc kubenswrapper[4733]: I1204 17:42:32.873472 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:42:33 crc kubenswrapper[4733]: I1204 17:42:33.090347 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:42:33 crc kubenswrapper[4733]: I1204 17:42:33.269423 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:42:33 crc kubenswrapper[4733]: I1204 17:42:33.269474 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:42:33 crc kubenswrapper[4733]: I1204 17:42:33.310198 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:42:34 crc kubenswrapper[4733]: I1204 17:42:34.089410 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:42:35 crc kubenswrapper[4733]: I1204 17:42:35.279537 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:42:35 crc kubenswrapper[4733]: I1204 17:42:35.280084 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:42:35 crc kubenswrapper[4733]: I1204 17:42:35.342041 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:42:35 crc kubenswrapper[4733]: I1204 17:42:35.968175 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cmjzf"] Dec 04 17:42:36 crc kubenswrapper[4733]: I1204 17:42:36.042614 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:42:36 crc kubenswrapper[4733]: I1204 17:42:36.042894 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:42:36 crc kubenswrapper[4733]: I1204 17:42:36.063095 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cmjzf" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerName="registry-server" containerID="cri-o://fbf7040307cd01072b5ab00eb0277eea0afac00461d5d9f9d4df8776050d576e" gracePeriod=2 Dec 04 17:42:36 crc kubenswrapper[4733]: I1204 17:42:36.115439 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:42:36 crc kubenswrapper[4733]: I1204 17:42:36.136943 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:42:36 crc kubenswrapper[4733]: I1204 17:42:36.426237 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:42:36 crc kubenswrapper[4733]: I1204 17:42:36.426308 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:42:36 crc kubenswrapper[4733]: I1204 17:42:36.493652 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:42:37 crc kubenswrapper[4733]: I1204 17:42:37.129284 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:42:37 crc kubenswrapper[4733]: I1204 17:42:37.135080 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:42:37 crc kubenswrapper[4733]: I1204 17:42:37.769389 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxkkc"] Dec 04 17:42:38 crc kubenswrapper[4733]: I1204 17:42:38.075896 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qxkkc" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="registry-server" containerID="cri-o://451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979" gracePeriod=2 Dec 04 17:42:39 crc kubenswrapper[4733]: I1204 17:42:39.082920 4733 generic.go:334] "Generic (PLEG): container finished" podID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerID="fbf7040307cd01072b5ab00eb0277eea0afac00461d5d9f9d4df8776050d576e" exitCode=0 Dec 04 17:42:39 crc kubenswrapper[4733]: I1204 17:42:39.083007 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmjzf" event={"ID":"3d1b51bf-b4c8-4a26-9be3-6c06322898db","Type":"ContainerDied","Data":"fbf7040307cd01072b5ab00eb0277eea0afac00461d5d9f9d4df8776050d576e"} Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.164216 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bxf9s"] Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.164476 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bxf9s" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="registry-server" containerID="cri-o://cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0" gracePeriod=2 Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.658214 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.815769 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx4wz\" (UniqueName: \"kubernetes.io/projected/3d1b51bf-b4c8-4a26-9be3-6c06322898db-kube-api-access-bx4wz\") pod \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.815876 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-utilities\") pod \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.815905 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-catalog-content\") pod \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\" (UID: \"3d1b51bf-b4c8-4a26-9be3-6c06322898db\") " Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.816927 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-utilities" (OuterVolumeSpecName: "utilities") pod "3d1b51bf-b4c8-4a26-9be3-6c06322898db" (UID: "3d1b51bf-b4c8-4a26-9be3-6c06322898db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.823002 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d1b51bf-b4c8-4a26-9be3-6c06322898db-kube-api-access-bx4wz" (OuterVolumeSpecName: "kube-api-access-bx4wz") pod "3d1b51bf-b4c8-4a26-9be3-6c06322898db" (UID: "3d1b51bf-b4c8-4a26-9be3-6c06322898db"). InnerVolumeSpecName "kube-api-access-bx4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.859567 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d1b51bf-b4c8-4a26-9be3-6c06322898db" (UID: "3d1b51bf-b4c8-4a26-9be3-6c06322898db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.917453 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.917501 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx4wz\" (UniqueName: \"kubernetes.io/projected/3d1b51bf-b4c8-4a26-9be3-6c06322898db-kube-api-access-bx4wz\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:40 crc kubenswrapper[4733]: I1204 17:42:40.917516 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d1b51bf-b4c8-4a26-9be3-6c06322898db-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.140343 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmjzf" event={"ID":"3d1b51bf-b4c8-4a26-9be3-6c06322898db","Type":"ContainerDied","Data":"d80a2dd23723c78826f12cc53561faa7cd7a9a24a50a0710c085329b4f526387"} Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.140369 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmjzf" Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.140401 4733 scope.go:117] "RemoveContainer" containerID="fbf7040307cd01072b5ab00eb0277eea0afac00461d5d9f9d4df8776050d576e" Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.143199 4733 generic.go:334] "Generic (PLEG): container finished" podID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerID="451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979" exitCode=0 Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.143222 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxkkc" event={"ID":"1b29e130-61cd-41e5-ab0b-e573e79d96cc","Type":"ContainerDied","Data":"451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979"} Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.160726 4733 scope.go:117] "RemoveContainer" containerID="9203b7f314f8ee7f803e63845c5dbc16979ec231a0d83fb4c57d4dde5b90733a" Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.174874 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cmjzf"] Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.177190 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cmjzf"] Dec 04 17:42:41 crc kubenswrapper[4733]: I1204 17:42:41.186304 4733 scope.go:117] "RemoveContainer" containerID="6e661c3146bbd2e2ee8b7946c86d37e94ad934b3a61ad0a8a921d85bb7f8f061" Dec 04 17:42:42 crc kubenswrapper[4733]: I1204 17:42:42.343256 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" path="/var/lib/kubelet/pods/3d1b51bf-b4c8-4a26-9be3-6c06322898db/volumes" Dec 04 17:42:45 crc kubenswrapper[4733]: E1204 17:42:45.280675 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979 is running failed: container process not found" containerID="451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 17:42:45 crc kubenswrapper[4733]: E1204 17:42:45.282053 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979 is running failed: container process not found" containerID="451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 17:42:45 crc kubenswrapper[4733]: E1204 17:42:45.282440 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979 is running failed: container process not found" containerID="451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 17:42:45 crc kubenswrapper[4733]: E1204 17:42:45.282477 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-qxkkc" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="registry-server" Dec 04 17:42:45 crc kubenswrapper[4733]: I1204 17:42:45.361823 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:42:45 crc kubenswrapper[4733]: I1204 17:42:45.361907 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:42:45 crc kubenswrapper[4733]: I1204 17:42:45.361977 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:42:45 crc kubenswrapper[4733]: I1204 17:42:45.362847 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 17:42:45 crc kubenswrapper[4733]: I1204 17:42:45.363037 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9" gracePeriod=600 Dec 04 17:42:46 crc kubenswrapper[4733]: E1204 17:42:46.427271 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0 is running failed: container process not found" containerID="cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 17:42:46 crc kubenswrapper[4733]: E1204 17:42:46.428001 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0 is running failed: container process not found" containerID="cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 17:42:46 crc kubenswrapper[4733]: E1204 17:42:46.428604 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0 is running failed: container process not found" containerID="cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 17:42:46 crc kubenswrapper[4733]: E1204 17:42:46.428691 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-bxf9s" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="registry-server" Dec 04 17:42:49 crc kubenswrapper[4733]: I1204 17:42:49.748974 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" podUID="ae851a3f-c43e-462e-b083-f8d2ba558f55" containerName="oauth-openshift" containerID="cri-o://9958d699a190321ffee4b80206d755daa1893938714280986a2983b478ad946d" gracePeriod=15 Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.296662 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bxf9s_2d42ea64-9dc6-4348-a5aa-f72c3f334775/registry-server/0.log" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.298103 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerID="cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0" exitCode=-1 Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.298144 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bxf9s" event={"ID":"2d42ea64-9dc6-4348-a5aa-f72c3f334775","Type":"ContainerDied","Data":"cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0"} Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.655958 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.750440 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvdrd\" (UniqueName: \"kubernetes.io/projected/1b29e130-61cd-41e5-ab0b-e573e79d96cc-kube-api-access-gvdrd\") pod \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.750689 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-utilities\") pod \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.750783 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-catalog-content\") pod \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\" (UID: \"1b29e130-61cd-41e5-ab0b-e573e79d96cc\") " Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.752366 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-utilities" (OuterVolumeSpecName: "utilities") pod "1b29e130-61cd-41e5-ab0b-e573e79d96cc" (UID: "1b29e130-61cd-41e5-ab0b-e573e79d96cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.770628 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b29e130-61cd-41e5-ab0b-e573e79d96cc-kube-api-access-gvdrd" (OuterVolumeSpecName: "kube-api-access-gvdrd") pod "1b29e130-61cd-41e5-ab0b-e573e79d96cc" (UID: "1b29e130-61cd-41e5-ab0b-e573e79d96cc"). InnerVolumeSpecName "kube-api-access-gvdrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.797618 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b29e130-61cd-41e5-ab0b-e573e79d96cc" (UID: "1b29e130-61cd-41e5-ab0b-e573e79d96cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.846030 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.852172 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.852205 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvdrd\" (UniqueName: \"kubernetes.io/projected/1b29e130-61cd-41e5-ab0b-e573e79d96cc-kube-api-access-gvdrd\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.852219 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b29e130-61cd-41e5-ab0b-e573e79d96cc-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.953276 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-catalog-content\") pod \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.953411 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl7t7\" (UniqueName: \"kubernetes.io/projected/2d42ea64-9dc6-4348-a5aa-f72c3f334775-kube-api-access-wl7t7\") pod \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.953441 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-utilities\") pod \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\" (UID: \"2d42ea64-9dc6-4348-a5aa-f72c3f334775\") " Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.954205 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-utilities" (OuterVolumeSpecName: "utilities") pod "2d42ea64-9dc6-4348-a5aa-f72c3f334775" (UID: "2d42ea64-9dc6-4348-a5aa-f72c3f334775"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:42:50 crc kubenswrapper[4733]: I1204 17:42:50.955937 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d42ea64-9dc6-4348-a5aa-f72c3f334775-kube-api-access-wl7t7" (OuterVolumeSpecName: "kube-api-access-wl7t7") pod "2d42ea64-9dc6-4348-a5aa-f72c3f334775" (UID: "2d42ea64-9dc6-4348-a5aa-f72c3f334775"). InnerVolumeSpecName "kube-api-access-wl7t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.055030 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl7t7\" (UniqueName: \"kubernetes.io/projected/2d42ea64-9dc6-4348-a5aa-f72c3f334775-kube-api-access-wl7t7\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.055073 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.057941 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d42ea64-9dc6-4348-a5aa-f72c3f334775" (UID: "2d42ea64-9dc6-4348-a5aa-f72c3f334775"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.156372 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d42ea64-9dc6-4348-a5aa-f72c3f334775-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.304876 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bxf9s" event={"ID":"2d42ea64-9dc6-4348-a5aa-f72c3f334775","Type":"ContainerDied","Data":"0aa1887f18131d19eca21fd98661cd5ced5db7a99017e2a3cd63055cd35eebb6"} Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.305317 4733 scope.go:117] "RemoveContainer" containerID="cc183b191b7915cd34edd63b36b5becb2e29b7bf1b098992f4af3f792d414eb0" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.305584 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bxf9s" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.309291 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxkkc" event={"ID":"1b29e130-61cd-41e5-ab0b-e573e79d96cc","Type":"ContainerDied","Data":"ee9a90a20acdd0b6bdbd38e088ba0569cf113321dc942ebb2432101e463b217e"} Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.309382 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxkkc" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.319078 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9" exitCode=0 Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.319131 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9"} Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.319154 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"adacb000dc8287db30045ce1678c37d9bee7575c1c73fa342ac484463a75e935"} Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.321902 4733 generic.go:334] "Generic (PLEG): container finished" podID="ae851a3f-c43e-462e-b083-f8d2ba558f55" containerID="9958d699a190321ffee4b80206d755daa1893938714280986a2983b478ad946d" exitCode=0 Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.322033 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" event={"ID":"ae851a3f-c43e-462e-b083-f8d2ba558f55","Type":"ContainerDied","Data":"9958d699a190321ffee4b80206d755daa1893938714280986a2983b478ad946d"} Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.323660 4733 scope.go:117] "RemoveContainer" containerID="610dba7a10b68f0f42dfefda016d20a12e9f91267d19568b6c19f924978fe4c6" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.353106 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxkkc"] Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.354734 4733 scope.go:117] "RemoveContainer" containerID="3054e1311a9fbb1200adfaac378f141053d07662b4037199d3d7b0d36080e682" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.357033 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxkkc"] Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.369992 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bxf9s"] Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.372155 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bxf9s"] Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.386139 4733 scope.go:117] "RemoveContainer" containerID="451aeb9878f895c195dbce5551eeb7ab3374264c36de635c75ac08ea25710979" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.397073 4733 scope.go:117] "RemoveContainer" containerID="026c36ed3b3eedfda502d93a5301bc0142cca40e78c8897ef45d2e2443259a61" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.407570 4733 scope.go:117] "RemoveContainer" containerID="f3d2e0fd1c792e22518b390173aaff1a3afeba2d30926e7575dd0390d40a40f4" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.439228 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.563601 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-service-ca\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.564309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.564399 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-cliconfig\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.564428 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhzrl\" (UniqueName: \"kubernetes.io/projected/ae851a3f-c43e-462e-b083-f8d2ba558f55-kube-api-access-lhzrl\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.564654 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.564709 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-provider-selection\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.564727 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-dir\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.564877 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565312 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-idp-0-file-data\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-trusted-ca-bundle\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565728 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-error\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565757 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-session\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565810 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-serving-cert\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565839 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-policies\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565872 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-ocp-branding-template\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565902 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-login\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.565926 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-router-certs\") pod \"ae851a3f-c43e-462e-b083-f8d2ba558f55\" (UID: \"ae851a3f-c43e-462e-b083-f8d2ba558f55\") " Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.566175 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.566630 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.566886 4733 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.566903 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.566912 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.566922 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.566932 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.569810 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae851a3f-c43e-462e-b083-f8d2ba558f55-kube-api-access-lhzrl" (OuterVolumeSpecName: "kube-api-access-lhzrl") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "kube-api-access-lhzrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.569819 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.569929 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.570144 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.570464 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.570461 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.570553 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.570627 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.570809 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ae851a3f-c43e-462e-b083-f8d2ba558f55" (UID: "ae851a3f-c43e-462e-b083-f8d2ba558f55"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.668483 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.668834 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.668989 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.669116 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.669235 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.669546 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhzrl\" (UniqueName: \"kubernetes.io/projected/ae851a3f-c43e-462e-b083-f8d2ba558f55-kube-api-access-lhzrl\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.669863 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.670012 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:51 crc kubenswrapper[4733]: I1204 17:42:51.670139 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ae851a3f-c43e-462e-b083-f8d2ba558f55-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:52 crc kubenswrapper[4733]: I1204 17:42:52.329322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" event={"ID":"ae851a3f-c43e-462e-b083-f8d2ba558f55","Type":"ContainerDied","Data":"d2b5de879c003f8ebbbb647156997fd37dcbe4469b9ce0c8b9bc137f5bb5e383"} Dec 04 17:42:52 crc kubenswrapper[4733]: I1204 17:42:52.330330 4733 scope.go:117] "RemoveContainer" containerID="9958d699a190321ffee4b80206d755daa1893938714280986a2983b478ad946d" Dec 04 17:42:52 crc kubenswrapper[4733]: I1204 17:42:52.329419 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bnzzn" Dec 04 17:42:52 crc kubenswrapper[4733]: I1204 17:42:52.345700 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" path="/var/lib/kubelet/pods/1b29e130-61cd-41e5-ab0b-e573e79d96cc/volumes" Dec 04 17:42:52 crc kubenswrapper[4733]: I1204 17:42:52.346814 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" path="/var/lib/kubelet/pods/2d42ea64-9dc6-4348-a5aa-f72c3f334775/volumes" Dec 04 17:42:52 crc kubenswrapper[4733]: I1204 17:42:52.360557 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bnzzn"] Dec 04 17:42:52 crc kubenswrapper[4733]: I1204 17:42:52.363239 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bnzzn"] Dec 04 17:42:54 crc kubenswrapper[4733]: I1204 17:42:54.342727 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae851a3f-c43e-462e-b083-f8d2ba558f55" path="/var/lib/kubelet/pods/ae851a3f-c43e-462e-b083-f8d2ba558f55/volumes" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.945986 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946564 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="extract-utilities" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946579 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="extract-utilities" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946589 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946597 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946612 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946620 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946630 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerName="extract-content" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946637 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerName="extract-content" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946647 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="extract-content" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946655 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="extract-content" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946666 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerName="extract-utilities" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946674 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerName="extract-utilities" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946684 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae851a3f-c43e-462e-b083-f8d2ba558f55" containerName="oauth-openshift" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946691 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae851a3f-c43e-462e-b083-f8d2ba558f55" containerName="oauth-openshift" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946703 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="extract-content" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946711 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="extract-content" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946723 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerName="extract-content" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946731 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerName="extract-content" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946743 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946751 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946762 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946769 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946779 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerName="extract-utilities" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946787 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerName="extract-utilities" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.946815 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="extract-utilities" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946824 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="extract-utilities" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946942 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbdd806-67e0-4af9-ab67-542319ad0d67" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946960 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d1b51bf-b4c8-4a26-9be3-6c06322898db" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946972 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae851a3f-c43e-462e-b083-f8d2ba558f55" containerName="oauth-openshift" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946985 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d42ea64-9dc6-4348-a5aa-f72c3f334775" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.946996 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b29e130-61cd-41e5-ab0b-e573e79d96cc" containerName="registry-server" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.947401 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.948323 4733 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.948726 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557" gracePeriod=15 Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.948811 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b" gracePeriod=15 Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.948774 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c" gracePeriod=15 Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.948904 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd" gracePeriod=15 Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.948930 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581" gracePeriod=15 Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.949979 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.950175 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.950194 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.950259 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.950268 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.950281 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.950289 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.950302 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.950310 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.950319 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.950326 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.950337 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.950344 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 17:42:56 crc kubenswrapper[4733]: E1204 17:42:56.950355 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.950363 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.951517 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.951538 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.951555 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.951576 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.951589 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 17:42:56 crc kubenswrapper[4733]: I1204 17:42:56.951904 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.010567 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.029990 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.030069 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.030096 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.030127 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.030149 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.030297 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.030348 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.030399 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.131649 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132190 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.131891 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132300 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132378 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132402 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132443 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132374 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132429 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132537 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132555 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132590 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132614 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132650 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132672 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.132783 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.305915 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:42:57 crc kubenswrapper[4733]: W1204 17:42:57.336300 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-77a15d0ad34205c0e36fe5cd4469e1e86bc40b5675e0dcd86d54d89cb7bdc872 WatchSource:0}: Error finding container 77a15d0ad34205c0e36fe5cd4469e1e86bc40b5675e0dcd86d54d89cb7bdc872: Status 404 returned error can't find the container with id 77a15d0ad34205c0e36fe5cd4469e1e86bc40b5675e0dcd86d54d89cb7bdc872 Dec 04 17:42:57 crc kubenswrapper[4733]: I1204 17:42:57.376617 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"77a15d0ad34205c0e36fe5cd4469e1e86bc40b5675e0dcd86d54d89cb7bdc872"} Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.031650 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.032114 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.342694 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.344205 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.387975 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.389337 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.390046 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c" exitCode=0 Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.390078 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581" exitCode=0 Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.390086 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b" exitCode=0 Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.390093 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd" exitCode=2 Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.390155 4733 scope.go:117] "RemoveContainer" containerID="e2343f5dfe0cbc6ed1cf48f0dbecd3f60ab1f078e0fcf9d2106c8bd974a83718" Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.391851 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c733fe32ebf56d635aea87ebbf29acc3ed1f572e04ad703fa96e507f1da89e8f"} Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.392834 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.394730 4733 generic.go:334] "Generic (PLEG): container finished" podID="dff4fe6a-a0a3-4832-800b-aa23959389bd" containerID="00c22ee848a6cc76bbd2ada8ccf6328f9493636f7be780ed8d3291c692f7ff0c" exitCode=0 Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.394766 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dff4fe6a-a0a3-4832-800b-aa23959389bd","Type":"ContainerDied","Data":"00c22ee848a6cc76bbd2ada8ccf6328f9493636f7be780ed8d3291c692f7ff0c"} Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.395385 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:42:58 crc kubenswrapper[4733]: I1204 17:42:58.395876 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:42:59 crc kubenswrapper[4733]: E1204 17:42:59.363489 4733 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.246:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" volumeName="registry-storage" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.402780 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.706851 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.707641 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.708604 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.868617 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-var-lock\") pod \"dff4fe6a-a0a3-4832-800b-aa23959389bd\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.869067 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dff4fe6a-a0a3-4832-800b-aa23959389bd-kube-api-access\") pod \"dff4fe6a-a0a3-4832-800b-aa23959389bd\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.869135 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-kubelet-dir\") pod \"dff4fe6a-a0a3-4832-800b-aa23959389bd\" (UID: \"dff4fe6a-a0a3-4832-800b-aa23959389bd\") " Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.868892 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-var-lock" (OuterVolumeSpecName: "var-lock") pod "dff4fe6a-a0a3-4832-800b-aa23959389bd" (UID: "dff4fe6a-a0a3-4832-800b-aa23959389bd"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.869442 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dff4fe6a-a0a3-4832-800b-aa23959389bd" (UID: "dff4fe6a-a0a3-4832-800b-aa23959389bd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.887770 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dff4fe6a-a0a3-4832-800b-aa23959389bd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dff4fe6a-a0a3-4832-800b-aa23959389bd" (UID: "dff4fe6a-a0a3-4832-800b-aa23959389bd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.970206 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.970242 4733 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dff4fe6a-a0a3-4832-800b-aa23959389bd-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 17:42:59 crc kubenswrapper[4733]: I1204 17:42:59.970252 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dff4fe6a-a0a3-4832-800b-aa23959389bd-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:00 crc kubenswrapper[4733]: I1204 17:43:00.413940 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dff4fe6a-a0a3-4832-800b-aa23959389bd","Type":"ContainerDied","Data":"6beb12c88f1b0e8bd0eea502dc1aaf57c69c5f4da7ee27c494a2c6b8e2b88c44"} Dec 04 17:43:00 crc kubenswrapper[4733]: I1204 17:43:00.413994 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6beb12c88f1b0e8bd0eea502dc1aaf57c69c5f4da7ee27c494a2c6b8e2b88c44" Dec 04 17:43:00 crc kubenswrapper[4733]: I1204 17:43:00.414133 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 17:43:00 crc kubenswrapper[4733]: I1204 17:43:00.420669 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:00 crc kubenswrapper[4733]: I1204 17:43:00.421547 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:01 crc kubenswrapper[4733]: E1204 17:43:01.126489 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:01 crc kubenswrapper[4733]: E1204 17:43:01.127363 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:01 crc kubenswrapper[4733]: E1204 17:43:01.127712 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:01 crc kubenswrapper[4733]: E1204 17:43:01.128247 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:01 crc kubenswrapper[4733]: E1204 17:43:01.128836 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:01 crc kubenswrapper[4733]: I1204 17:43:01.128901 4733 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 04 17:43:01 crc kubenswrapper[4733]: E1204 17:43:01.129339 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="200ms" Dec 04 17:43:01 crc kubenswrapper[4733]: E1204 17:43:01.330533 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="400ms" Dec 04 17:43:01 crc kubenswrapper[4733]: E1204 17:43:01.732121 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="800ms" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.436227 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.437437 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557" exitCode=0 Dec 04 17:43:02 crc kubenswrapper[4733]: E1204 17:43:02.533620 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="1.6s" Dec 04 17:43:02 crc kubenswrapper[4733]: E1204 17:43:02.789869 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.246:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e14137586bb07 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 17:42:57.784511239 +0000 UTC m=+239.739872325,LastTimestamp:2025-12-04 17:42:57.784511239 +0000 UTC m=+239.739872325,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.803478 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.804269 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.804730 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.805040 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.805346 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912116 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912182 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912237 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912245 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912281 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912401 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912654 4733 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912676 4733 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:02 crc kubenswrapper[4733]: I1204 17:43:02.912686 4733 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.447494 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.448985 4733 scope.go:117] "RemoveContainer" containerID="0ed83f09bf4a5072141d616cf2cfe39540962e4cbd2afcc2918f52de79e6496c" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.449063 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.469617 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.470875 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.471674 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.475108 4733 scope.go:117] "RemoveContainer" containerID="05885c1f6ed92f8641a05c2933a2ca55bf3226eecf975324a2a3ef3af2fc5581" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.498066 4733 scope.go:117] "RemoveContainer" containerID="29634f33bc7f0aa4b3cce6fc3fc0eed09098b88e1b1a036400fdb9337552fb1b" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.517017 4733 scope.go:117] "RemoveContainer" containerID="617a45d1725945014bc39d39247cf3d82ba89e8b9a9d34f61952979c8d51d8bd" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.540374 4733 scope.go:117] "RemoveContainer" containerID="746d831a60131bbb7f5948fc90a5204eab13e44785da683bdad68dff8d582557" Dec 04 17:43:03 crc kubenswrapper[4733]: I1204 17:43:03.563995 4733 scope.go:117] "RemoveContainer" containerID="fa199e908023da40b383ab8b78c92bb2b0e2b810f0357d06a7ae8fef214f43a7" Dec 04 17:43:04 crc kubenswrapper[4733]: E1204 17:43:04.007974 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:43:04Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:43:04Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:43:04Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T17:43:04Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:04 crc kubenswrapper[4733]: E1204 17:43:04.008677 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:04 crc kubenswrapper[4733]: E1204 17:43:04.009168 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:04 crc kubenswrapper[4733]: E1204 17:43:04.009559 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:04 crc kubenswrapper[4733]: E1204 17:43:04.009921 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:04 crc kubenswrapper[4733]: E1204 17:43:04.009959 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 17:43:04 crc kubenswrapper[4733]: E1204 17:43:04.134935 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="3.2s" Dec 04 17:43:04 crc kubenswrapper[4733]: I1204 17:43:04.349581 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 04 17:43:05 crc kubenswrapper[4733]: E1204 17:43:05.677699 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.246:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e14137586bb07 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 17:42:57.784511239 +0000 UTC m=+239.739872325,LastTimestamp:2025-12-04 17:42:57.784511239 +0000 UTC m=+239.739872325,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 17:43:07 crc kubenswrapper[4733]: E1204 17:43:07.335592 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="6.4s" Dec 04 17:43:08 crc kubenswrapper[4733]: I1204 17:43:08.339520 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:08 crc kubenswrapper[4733]: I1204 17:43:08.340514 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:11 crc kubenswrapper[4733]: I1204 17:43:11.334747 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:11 crc kubenswrapper[4733]: I1204 17:43:11.335713 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:11 crc kubenswrapper[4733]: I1204 17:43:11.335944 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:11 crc kubenswrapper[4733]: I1204 17:43:11.395852 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:11 crc kubenswrapper[4733]: I1204 17:43:11.396238 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:11 crc kubenswrapper[4733]: E1204 17:43:11.396822 4733 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:11 crc kubenswrapper[4733]: I1204 17:43:11.397289 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:11 crc kubenswrapper[4733]: I1204 17:43:11.509393 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"86fa5b9936df3edcb64adda4c78e136d9d017157042377efbe0fd7b2f59e12ae"} Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.521070 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.521416 4733 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45" exitCode=1 Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.521501 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45"} Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.522153 4733 scope.go:117] "RemoveContainer" containerID="b391edf90fa409e2601e4535163830d679e8ac838a0641b2ebb54267c6c70b45" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.523228 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.524102 4733 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.524879 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.525331 4733 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="31970ed6937535c5e0e329063f8b85650e5c48c7ee0e47941c2e6f1a394b3755" exitCode=0 Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.525374 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"31970ed6937535c5e0e329063f8b85650e5c48c7ee0e47941c2e6f1a394b3755"} Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.525774 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.525834 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:12 crc kubenswrapper[4733]: E1204 17:43:12.526243 4733 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.526244 4733 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.526892 4733 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:12 crc kubenswrapper[4733]: I1204 17:43:12.527239 4733 status_manager.go:851] "Failed to get status for pod" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Dec 04 17:43:13 crc kubenswrapper[4733]: I1204 17:43:13.227376 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:43:13 crc kubenswrapper[4733]: I1204 17:43:13.540617 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 17:43:13 crc kubenswrapper[4733]: I1204 17:43:13.540777 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b923409b694ac2311a62ce154c58d547a3093dc27783a7e6f1eee0677305d395"} Dec 04 17:43:13 crc kubenswrapper[4733]: I1204 17:43:13.546367 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b3e7716de2b618d8dcfc0c9a7a4d8590af4b77e67028d426d5581d544e17dcc2"} Dec 04 17:43:13 crc kubenswrapper[4733]: I1204 17:43:13.546407 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2d3b211e2986e2a0eafe4780e06c7871261cc42824f8be8b533c31c63057fea7"} Dec 04 17:43:13 crc kubenswrapper[4733]: I1204 17:43:13.546421 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e0469a889f8030970f5adf2bb75cb85c48991e4feefca6d568b0c128eff1999e"} Dec 04 17:43:13 crc kubenswrapper[4733]: I1204 17:43:13.546445 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4c65e5a3e68ff9ffa6cf1cf29c4b235c62088e5a1842dfe84bd3f18fc56419f9"} Dec 04 17:43:14 crc kubenswrapper[4733]: I1204 17:43:14.554413 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cc3b9d8d3756d1c29d60b824c8aa016b1a4843048fca213be0abbb590024c4df"} Dec 04 17:43:14 crc kubenswrapper[4733]: I1204 17:43:14.554697 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:14 crc kubenswrapper[4733]: I1204 17:43:14.554857 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:16 crc kubenswrapper[4733]: I1204 17:43:16.398176 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:16 crc kubenswrapper[4733]: I1204 17:43:16.398619 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:16 crc kubenswrapper[4733]: I1204 17:43:16.405064 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:18 crc kubenswrapper[4733]: I1204 17:43:18.033147 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:43:19 crc kubenswrapper[4733]: I1204 17:43:19.562659 4733 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:19 crc kubenswrapper[4733]: I1204 17:43:19.584842 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:19 crc kubenswrapper[4733]: I1204 17:43:19.584856 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:19 crc kubenswrapper[4733]: I1204 17:43:19.585179 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:19 crc kubenswrapper[4733]: I1204 17:43:19.588784 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:19 crc kubenswrapper[4733]: I1204 17:43:19.591238 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4c09d343-1a94-4a6f-8a1f-4b40c75e0463" Dec 04 17:43:20 crc kubenswrapper[4733]: I1204 17:43:20.591921 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:20 crc kubenswrapper[4733]: I1204 17:43:20.591971 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:20 crc kubenswrapper[4733]: I1204 17:43:20.596428 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4c09d343-1a94-4a6f-8a1f-4b40c75e0463" Dec 04 17:43:21 crc kubenswrapper[4733]: I1204 17:43:21.598167 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:21 crc kubenswrapper[4733]: I1204 17:43:21.598502 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:21 crc kubenswrapper[4733]: I1204 17:43:21.602472 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4c09d343-1a94-4a6f-8a1f-4b40c75e0463" Dec 04 17:43:23 crc kubenswrapper[4733]: I1204 17:43:23.227668 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:43:23 crc kubenswrapper[4733]: I1204 17:43:23.235813 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:43:23 crc kubenswrapper[4733]: I1204 17:43:23.612528 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 17:43:30 crc kubenswrapper[4733]: I1204 17:43:30.236467 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 17:43:30 crc kubenswrapper[4733]: I1204 17:43:30.270034 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 17:43:30 crc kubenswrapper[4733]: I1204 17:43:30.374566 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 17:43:30 crc kubenswrapper[4733]: I1204 17:43:30.663046 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 17:43:30 crc kubenswrapper[4733]: I1204 17:43:30.759372 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 17:43:30 crc kubenswrapper[4733]: I1204 17:43:30.982481 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.077028 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.122263 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.224361 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.300534 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.504403 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.514623 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.630319 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.731122 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.945573 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 17:43:31 crc kubenswrapper[4733]: I1204 17:43:31.956145 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.226179 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.349771 4733 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.440719 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.533161 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.534701 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.535950 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.750552 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.770654 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.848722 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.850232 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.868111 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.912297 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 17:43:32 crc kubenswrapper[4733]: I1204 17:43:32.999688 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.093341 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.203851 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.221254 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.365736 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.386687 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.404081 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.493103 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.493285 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.630995 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.731870 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.809446 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 17:43:33 crc kubenswrapper[4733]: I1204 17:43:33.851295 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.091722 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.098857 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.132937 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.166732 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.219653 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.263258 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.272350 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.341173 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.398330 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.461964 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.548333 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.626721 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.631386 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.659955 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.732169 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.769082 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.782131 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.829664 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.881772 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.886388 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.932239 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 17:43:34 crc kubenswrapper[4733]: I1204 17:43:34.997030 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.034083 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.108956 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.176129 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.176252 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.276430 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.308349 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.390935 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.459966 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.548321 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.584877 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.646915 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.715317 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.766531 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.777416 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.828848 4733 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.833181 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 17:43:35 crc kubenswrapper[4733]: I1204 17:43:35.964243 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.001301 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.158166 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.230389 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.246195 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.276106 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.424470 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.495210 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.558151 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.611432 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.797494 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.831091 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.881084 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.907995 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 17:43:36 crc kubenswrapper[4733]: I1204 17:43:36.922248 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.026116 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.140600 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.175432 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.183184 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.243079 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.273988 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.314925 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.339042 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.396549 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.443975 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.496088 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.528959 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.532343 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.642153 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.653752 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.703729 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.729407 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.743126 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.744611 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.776552 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.850668 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.889303 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 17:43:37 crc kubenswrapper[4733]: I1204 17:43:37.972007 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.022044 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.049372 4733 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.053849 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=42.053826931 podStartE2EDuration="42.053826931s" podCreationTimestamp="2025-12-04 17:42:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:43:19.506920614 +0000 UTC m=+261.462281670" watchObservedRunningTime="2025-12-04 17:43:38.053826931 +0000 UTC m=+280.009188017" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.057224 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.058266 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-6f94556d49-km6h7"] Dec 04 17:43:38 crc kubenswrapper[4733]: E1204 17:43:38.059059 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" containerName="installer" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.059103 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" containerName="installer" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.059295 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="dff4fe6a-a0a3-4832-800b-aa23959389bd" containerName="installer" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.060022 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.061425 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.061467 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="73cceba9-c503-4d25-8c75-b634e1bb59d0" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.064281 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.064627 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.064705 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.064759 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.067042 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.067259 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.067374 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.067566 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.067575 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.067439 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.067622 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.067490 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.068220 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.083325 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.085178 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.095040 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.104580 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.104555162 podStartE2EDuration="19.104555162s" podCreationTimestamp="2025-12-04 17:43:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:43:38.100377046 +0000 UTC m=+280.055738102" watchObservedRunningTime="2025-12-04 17:43:38.104555162 +0000 UTC m=+280.059916248" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.203642 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-session\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.203699 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.203734 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-login\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.203764 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.203822 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.203847 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.203878 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-audit-dir\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.203907 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vckrr\" (UniqueName: \"kubernetes.io/projected/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-kube-api-access-vckrr\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.204021 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.204096 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-audit-policies\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.204131 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.204161 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-error\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.204192 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.204219 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305107 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-error\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305192 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305228 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305277 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-session\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305326 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305371 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-login\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305411 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305465 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.305616 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.306171 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-audit-dir\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.306526 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vckrr\" (UniqueName: \"kubernetes.io/projected/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-kube-api-access-vckrr\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.306558 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.306587 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-audit-policies\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.306609 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.306713 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.306872 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-audit-dir\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.306976 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.308241 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-audit-policies\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.308432 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.312004 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.312162 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.312293 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-session\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.313203 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-error\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.314045 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.315326 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.316294 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-template-login\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.328873 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.340170 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.363943 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vckrr\" (UniqueName: \"kubernetes.io/projected/889a62e0-ac6c-464c-bc23-5e6bd4096dbf-kube-api-access-vckrr\") pod \"oauth-openshift-6f94556d49-km6h7\" (UID: \"889a62e0-ac6c-464c-bc23-5e6bd4096dbf\") " pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.386233 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.393870 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.403426 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.406715 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.445359 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.499814 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.527537 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.592353 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.624093 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.634249 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.642519 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.800869 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.866895 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.886747 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.913699 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.926643 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.932539 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 17:43:38 crc kubenswrapper[4733]: I1204 17:43:38.988947 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.012017 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.196941 4733 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.199620 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.263873 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.364957 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.366995 4733 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.379469 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.516787 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.540965 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.550074 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.587787 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.591418 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.600388 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.642845 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.657155 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.658221 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.683633 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.688271 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.763926 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.764226 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.834386 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.845865 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 17:43:39 crc kubenswrapper[4733]: I1204 17:43:39.847740 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.081966 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.189770 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.199767 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.574396 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.652624 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.657008 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.672219 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.728408 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.776002 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.804755 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.827780 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.907677 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.933115 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 17:43:40 crc kubenswrapper[4733]: I1204 17:43:40.965674 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.058382 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.210237 4733 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.238280 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.308285 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.321061 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.358243 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.392437 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.402085 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 17:43:41 crc kubenswrapper[4733]: E1204 17:43:41.408623 4733 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 04 17:43:41 crc kubenswrapper[4733]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f94556d49-km6h7_openshift-authentication_889a62e0-ac6c-464c-bc23-5e6bd4096dbf_0(0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d): error adding pod openshift-authentication_oauth-openshift-6f94556d49-km6h7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d" Netns:"/var/run/netns/b724c658-e462-4e17-9c83-6dd9ae473a95" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f94556d49-km6h7;K8S_POD_INFRA_CONTAINER_ID=0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d;K8S_POD_UID=889a62e0-ac6c-464c-bc23-5e6bd4096dbf" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f94556d49-km6h7] networking: Multus: [openshift-authentication/oauth-openshift-6f94556d49-km6h7/889a62e0-ac6c-464c-bc23-5e6bd4096dbf]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f94556d49-km6h7 in out of cluster comm: pod "oauth-openshift-6f94556d49-km6h7" not found Dec 04 17:43:41 crc kubenswrapper[4733]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 17:43:41 crc kubenswrapper[4733]: > Dec 04 17:43:41 crc kubenswrapper[4733]: E1204 17:43:41.408702 4733 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 04 17:43:41 crc kubenswrapper[4733]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f94556d49-km6h7_openshift-authentication_889a62e0-ac6c-464c-bc23-5e6bd4096dbf_0(0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d): error adding pod openshift-authentication_oauth-openshift-6f94556d49-km6h7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d" Netns:"/var/run/netns/b724c658-e462-4e17-9c83-6dd9ae473a95" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f94556d49-km6h7;K8S_POD_INFRA_CONTAINER_ID=0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d;K8S_POD_UID=889a62e0-ac6c-464c-bc23-5e6bd4096dbf" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f94556d49-km6h7] networking: Multus: [openshift-authentication/oauth-openshift-6f94556d49-km6h7/889a62e0-ac6c-464c-bc23-5e6bd4096dbf]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f94556d49-km6h7 in out of cluster comm: pod "oauth-openshift-6f94556d49-km6h7" not found Dec 04 17:43:41 crc kubenswrapper[4733]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 17:43:41 crc kubenswrapper[4733]: > pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:41 crc kubenswrapper[4733]: E1204 17:43:41.408726 4733 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 04 17:43:41 crc kubenswrapper[4733]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f94556d49-km6h7_openshift-authentication_889a62e0-ac6c-464c-bc23-5e6bd4096dbf_0(0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d): error adding pod openshift-authentication_oauth-openshift-6f94556d49-km6h7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d" Netns:"/var/run/netns/b724c658-e462-4e17-9c83-6dd9ae473a95" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f94556d49-km6h7;K8S_POD_INFRA_CONTAINER_ID=0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d;K8S_POD_UID=889a62e0-ac6c-464c-bc23-5e6bd4096dbf" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f94556d49-km6h7] networking: Multus: [openshift-authentication/oauth-openshift-6f94556d49-km6h7/889a62e0-ac6c-464c-bc23-5e6bd4096dbf]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f94556d49-km6h7 in out of cluster comm: pod "oauth-openshift-6f94556d49-km6h7" not found Dec 04 17:43:41 crc kubenswrapper[4733]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 17:43:41 crc kubenswrapper[4733]: > pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:41 crc kubenswrapper[4733]: E1204 17:43:41.408783 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-6f94556d49-km6h7_openshift-authentication(889a62e0-ac6c-464c-bc23-5e6bd4096dbf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-6f94556d49-km6h7_openshift-authentication(889a62e0-ac6c-464c-bc23-5e6bd4096dbf)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f94556d49-km6h7_openshift-authentication_889a62e0-ac6c-464c-bc23-5e6bd4096dbf_0(0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d): error adding pod openshift-authentication_oauth-openshift-6f94556d49-km6h7 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d\\\" Netns:\\\"/var/run/netns/b724c658-e462-4e17-9c83-6dd9ae473a95\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f94556d49-km6h7;K8S_POD_INFRA_CONTAINER_ID=0188352570a9e046f0c5a39ba32ee70d79a82e1db011cfe0b37bd224dbcead7d;K8S_POD_UID=889a62e0-ac6c-464c-bc23-5e6bd4096dbf\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f94556d49-km6h7] networking: Multus: [openshift-authentication/oauth-openshift-6f94556d49-km6h7/889a62e0-ac6c-464c-bc23-5e6bd4096dbf]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f94556d49-km6h7 in out of cluster comm: pod \\\"oauth-openshift-6f94556d49-km6h7\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" podUID="889a62e0-ac6c-464c-bc23-5e6bd4096dbf" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.445901 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.446101 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.462652 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.525920 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.715683 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.755203 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.889740 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 17:43:41 crc kubenswrapper[4733]: I1204 17:43:41.925548 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.000188 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.082299 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.183638 4733 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.184014 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://c733fe32ebf56d635aea87ebbf29acc3ed1f572e04ad703fa96e507f1da89e8f" gracePeriod=5 Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.418700 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.537099 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.648053 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.800074 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.802323 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.842774 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.883529 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.907613 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 17:43:42 crc kubenswrapper[4733]: I1204 17:43:42.968002 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.058214 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.072371 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.116705 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.165070 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.263820 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.277126 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.347049 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.347110 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.400127 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.482276 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.512449 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.557715 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.777761 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.800406 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 17:43:43 crc kubenswrapper[4733]: I1204 17:43:43.933057 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.145928 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.192364 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.198787 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.236896 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.281129 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.305945 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.362017 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.426932 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.529461 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.549937 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.719910 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.923194 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 17:43:44 crc kubenswrapper[4733]: I1204 17:43:44.983095 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 17:43:45 crc kubenswrapper[4733]: I1204 17:43:45.018703 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 17:43:45 crc kubenswrapper[4733]: I1204 17:43:45.186734 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 17:43:45 crc kubenswrapper[4733]: I1204 17:43:45.387927 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 17:43:45 crc kubenswrapper[4733]: I1204 17:43:45.421384 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 17:43:45 crc kubenswrapper[4733]: I1204 17:43:45.431409 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 17:43:45 crc kubenswrapper[4733]: I1204 17:43:45.527104 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 17:43:46 crc kubenswrapper[4733]: I1204 17:43:46.269951 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 17:43:46 crc kubenswrapper[4733]: I1204 17:43:46.699042 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.780116 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.780245 4733 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="c733fe32ebf56d635aea87ebbf29acc3ed1f572e04ad703fa96e507f1da89e8f" exitCode=137 Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.780323 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77a15d0ad34205c0e36fe5cd4469e1e86bc40b5675e0dcd86d54d89cb7bdc872" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.789075 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.789181 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944143 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944234 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944294 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944386 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944445 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944467 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944455 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944539 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.944603 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.945028 4733 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.945069 4733 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.945090 4733 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.945108 4733 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:47 crc kubenswrapper[4733]: I1204 17:43:47.958304 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:43:48 crc kubenswrapper[4733]: I1204 17:43:48.046265 4733 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 17:43:48 crc kubenswrapper[4733]: I1204 17:43:48.344748 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 04 17:43:48 crc kubenswrapper[4733]: I1204 17:43:48.345631 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 04 17:43:48 crc kubenswrapper[4733]: I1204 17:43:48.359259 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 17:43:48 crc kubenswrapper[4733]: I1204 17:43:48.359594 4733 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="79b08751-ac52-448e-8cd9-29dc13cb54e1" Dec 04 17:43:48 crc kubenswrapper[4733]: I1204 17:43:48.363816 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 17:43:48 crc kubenswrapper[4733]: I1204 17:43:48.363862 4733 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="79b08751-ac52-448e-8cd9-29dc13cb54e1" Dec 04 17:43:48 crc kubenswrapper[4733]: I1204 17:43:48.786840 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 17:43:56 crc kubenswrapper[4733]: I1204 17:43:56.335304 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:56 crc kubenswrapper[4733]: I1204 17:43:56.336937 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:56 crc kubenswrapper[4733]: I1204 17:43:56.781500 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6f94556d49-km6h7"] Dec 04 17:43:56 crc kubenswrapper[4733]: I1204 17:43:56.842885 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" event={"ID":"889a62e0-ac6c-464c-bc23-5e6bd4096dbf","Type":"ContainerStarted","Data":"3551880e0e67c571d31ccc306150fca7dce125c6669fa2648c41eef7e8b0c0d4"} Dec 04 17:43:57 crc kubenswrapper[4733]: I1204 17:43:57.852565 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" event={"ID":"889a62e0-ac6c-464c-bc23-5e6bd4096dbf","Type":"ContainerStarted","Data":"fab87a9193090d8266db843479d291c3c457b18f5cd944e7baee3c7e6ab243c6"} Dec 04 17:43:57 crc kubenswrapper[4733]: I1204 17:43:57.853150 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:57 crc kubenswrapper[4733]: I1204 17:43:57.863920 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" Dec 04 17:43:57 crc kubenswrapper[4733]: I1204 17:43:57.882065 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6f94556d49-km6h7" podStartSLOduration=93.882043377 podStartE2EDuration="1m33.882043377s" podCreationTimestamp="2025-12-04 17:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:43:57.880871865 +0000 UTC m=+299.836232961" watchObservedRunningTime="2025-12-04 17:43:57.882043377 +0000 UTC m=+299.837404433" Dec 04 17:44:00 crc kubenswrapper[4733]: I1204 17:44:00.876528 4733 generic.go:334] "Generic (PLEG): container finished" podID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerID="5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3" exitCode=0 Dec 04 17:44:00 crc kubenswrapper[4733]: I1204 17:44:00.876659 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" event={"ID":"7caf3411-fca7-4025-b9b6-eb3b4a43ab81","Type":"ContainerDied","Data":"5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3"} Dec 04 17:44:00 crc kubenswrapper[4733]: I1204 17:44:00.877680 4733 scope.go:117] "RemoveContainer" containerID="5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3" Dec 04 17:44:01 crc kubenswrapper[4733]: I1204 17:44:01.886892 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" event={"ID":"7caf3411-fca7-4025-b9b6-eb3b4a43ab81","Type":"ContainerStarted","Data":"2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526"} Dec 04 17:44:01 crc kubenswrapper[4733]: I1204 17:44:01.887826 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:44:01 crc kubenswrapper[4733]: I1204 17:44:01.890149 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.179776 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8j2hf"] Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.180641 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" podUID="d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" containerName="controller-manager" containerID="cri-o://ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637" gracePeriod=30 Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.275870 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92"] Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.276494 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" podUID="8e4a43e7-58da-4eb1-930f-1350cde7398b" containerName="route-controller-manager" containerID="cri-o://a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8" gracePeriod=30 Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.569937 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.618474 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753074 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95nrs\" (UniqueName: \"kubernetes.io/projected/8e4a43e7-58da-4eb1-930f-1350cde7398b-kube-api-access-95nrs\") pod \"8e4a43e7-58da-4eb1-930f-1350cde7398b\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753137 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-client-ca\") pod \"8e4a43e7-58da-4eb1-930f-1350cde7398b\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753231 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgckf\" (UniqueName: \"kubernetes.io/projected/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-kube-api-access-xgckf\") pod \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753271 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-serving-cert\") pod \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753340 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e4a43e7-58da-4eb1-930f-1350cde7398b-serving-cert\") pod \"8e4a43e7-58da-4eb1-930f-1350cde7398b\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753387 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-client-ca\") pod \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753423 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-proxy-ca-bundles\") pod \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753453 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-config\") pod \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\" (UID: \"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.753487 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-config\") pod \"8e4a43e7-58da-4eb1-930f-1350cde7398b\" (UID: \"8e4a43e7-58da-4eb1-930f-1350cde7398b\") " Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754244 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" (UID: "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754411 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-config" (OuterVolumeSpecName: "config") pod "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" (UID: "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754472 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-client-ca" (OuterVolumeSpecName: "client-ca") pod "8e4a43e7-58da-4eb1-930f-1350cde7398b" (UID: "8e4a43e7-58da-4eb1-930f-1350cde7398b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754493 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" (UID: "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754533 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-config" (OuterVolumeSpecName: "config") pod "8e4a43e7-58da-4eb1-930f-1350cde7398b" (UID: "8e4a43e7-58da-4eb1-930f-1350cde7398b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754893 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754918 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754929 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754939 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.754953 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e4a43e7-58da-4eb1-930f-1350cde7398b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.760445 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-kube-api-access-xgckf" (OuterVolumeSpecName: "kube-api-access-xgckf") pod "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" (UID: "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7"). InnerVolumeSpecName "kube-api-access-xgckf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.760658 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e4a43e7-58da-4eb1-930f-1350cde7398b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8e4a43e7-58da-4eb1-930f-1350cde7398b" (UID: "8e4a43e7-58da-4eb1-930f-1350cde7398b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.761030 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" (UID: "d6ba934e-90ea-47d0-9c7f-7ceca992d5b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.761080 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e4a43e7-58da-4eb1-930f-1350cde7398b-kube-api-access-95nrs" (OuterVolumeSpecName: "kube-api-access-95nrs") pod "8e4a43e7-58da-4eb1-930f-1350cde7398b" (UID: "8e4a43e7-58da-4eb1-930f-1350cde7398b"). InnerVolumeSpecName "kube-api-access-95nrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.856124 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95nrs\" (UniqueName: \"kubernetes.io/projected/8e4a43e7-58da-4eb1-930f-1350cde7398b-kube-api-access-95nrs\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.856172 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgckf\" (UniqueName: \"kubernetes.io/projected/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-kube-api-access-xgckf\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.856194 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.856213 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e4a43e7-58da-4eb1-930f-1350cde7398b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.989524 4733 generic.go:334] "Generic (PLEG): container finished" podID="d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" containerID="ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637" exitCode=0 Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.989644 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" event={"ID":"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7","Type":"ContainerDied","Data":"ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637"} Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.989650 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.989690 4733 scope.go:117] "RemoveContainer" containerID="ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.989677 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8j2hf" event={"ID":"d6ba934e-90ea-47d0-9c7f-7ceca992d5b7","Type":"ContainerDied","Data":"cbad3c1658fbbf362bfc71e73e1417d6aabeb88bc2cc80c59bf823f02c83879e"} Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.992343 4733 generic.go:334] "Generic (PLEG): container finished" podID="8e4a43e7-58da-4eb1-930f-1350cde7398b" containerID="a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8" exitCode=0 Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.992382 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" event={"ID":"8e4a43e7-58da-4eb1-930f-1350cde7398b","Type":"ContainerDied","Data":"a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8"} Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.992416 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" Dec 04 17:44:17 crc kubenswrapper[4733]: I1204 17:44:17.992586 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92" event={"ID":"8e4a43e7-58da-4eb1-930f-1350cde7398b","Type":"ContainerDied","Data":"0b08c4d845f60f35971aaf4a7fcd9f73668939933d7b7db9f7f22eaaebb28273"} Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.013148 4733 scope.go:117] "RemoveContainer" containerID="ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637" Dec 04 17:44:18 crc kubenswrapper[4733]: E1204 17:44:18.014144 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637\": container with ID starting with ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637 not found: ID does not exist" containerID="ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.014194 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637"} err="failed to get container status \"ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637\": rpc error: code = NotFound desc = could not find container \"ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637\": container with ID starting with ea7fe7afaabfd33edcaa5a461d203deeaff9b0b899a4bbc2f578a9d57dc54637 not found: ID does not exist" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.014258 4733 scope.go:117] "RemoveContainer" containerID="a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.041203 4733 scope.go:117] "RemoveContainer" containerID="a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8" Dec 04 17:44:18 crc kubenswrapper[4733]: E1204 17:44:18.042247 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8\": container with ID starting with a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8 not found: ID does not exist" containerID="a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.042334 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8"} err="failed to get container status \"a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8\": rpc error: code = NotFound desc = could not find container \"a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8\": container with ID starting with a1504572d3eecc55482af675737e3c83fd7e6ce40e6e8f6f21b676d545d5b7f8 not found: ID does not exist" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.046641 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92"] Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.056018 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vbf92"] Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.063681 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8j2hf"] Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.070898 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8j2hf"] Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.350748 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e4a43e7-58da-4eb1-930f-1350cde7398b" path="/var/lib/kubelet/pods/8e4a43e7-58da-4eb1-930f-1350cde7398b/volumes" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.352378 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" path="/var/lib/kubelet/pods/d6ba934e-90ea-47d0-9c7f-7ceca992d5b7/volumes" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.734681 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm"] Dec 04 17:44:18 crc kubenswrapper[4733]: E1204 17:44:18.735054 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" containerName="controller-manager" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.735085 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" containerName="controller-manager" Dec 04 17:44:18 crc kubenswrapper[4733]: E1204 17:44:18.735109 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4a43e7-58da-4eb1-930f-1350cde7398b" containerName="route-controller-manager" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.735124 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4a43e7-58da-4eb1-930f-1350cde7398b" containerName="route-controller-manager" Dec 04 17:44:18 crc kubenswrapper[4733]: E1204 17:44:18.735155 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.735170 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.735369 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.735400 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6ba934e-90ea-47d0-9c7f-7ceca992d5b7" containerName="controller-manager" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.735420 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e4a43e7-58da-4eb1-930f-1350cde7398b" containerName="route-controller-manager" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.736153 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.744234 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.744589 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.746195 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.746337 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.746524 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.746755 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.755954 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-85b9785f75-r4tkq"] Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.764372 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.767856 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.768389 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.768590 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.768788 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85b9785f75-r4tkq"] Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.771029 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.771203 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.772506 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.786021 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm"] Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.792062 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868462 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-config\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868529 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-client-ca\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868572 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-proxy-ca-bundles\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868708 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a512712c-0f23-47a9-8293-f66628fa3605-config\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868757 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-624lv\" (UniqueName: \"kubernetes.io/projected/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-kube-api-access-624lv\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868822 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-serving-cert\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868855 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a512712c-0f23-47a9-8293-f66628fa3605-client-ca\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868908 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a512712c-0f23-47a9-8293-f66628fa3605-serving-cert\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.868944 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kp7x\" (UniqueName: \"kubernetes.io/projected/a512712c-0f23-47a9-8293-f66628fa3605-kube-api-access-7kp7x\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.970542 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-config\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.970618 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-client-ca\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.970669 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-proxy-ca-bundles\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.970764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-624lv\" (UniqueName: \"kubernetes.io/projected/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-kube-api-access-624lv\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.970829 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a512712c-0f23-47a9-8293-f66628fa3605-config\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.970862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-serving-cert\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.970896 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a512712c-0f23-47a9-8293-f66628fa3605-client-ca\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.971003 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a512712c-0f23-47a9-8293-f66628fa3605-serving-cert\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.971037 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kp7x\" (UniqueName: \"kubernetes.io/projected/a512712c-0f23-47a9-8293-f66628fa3605-kube-api-access-7kp7x\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.972372 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-client-ca\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.972985 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-proxy-ca-bundles\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.973356 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a512712c-0f23-47a9-8293-f66628fa3605-client-ca\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.974647 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-config\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.976561 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a512712c-0f23-47a9-8293-f66628fa3605-config\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.976742 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-serving-cert\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.978223 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a512712c-0f23-47a9-8293-f66628fa3605-serving-cert\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.998722 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kp7x\" (UniqueName: \"kubernetes.io/projected/a512712c-0f23-47a9-8293-f66628fa3605-kube-api-access-7kp7x\") pod \"route-controller-manager-5f564968d7-fngcm\" (UID: \"a512712c-0f23-47a9-8293-f66628fa3605\") " pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:18 crc kubenswrapper[4733]: I1204 17:44:18.999581 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-624lv\" (UniqueName: \"kubernetes.io/projected/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-kube-api-access-624lv\") pod \"controller-manager-85b9785f75-r4tkq\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:19 crc kubenswrapper[4733]: I1204 17:44:19.078453 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:19 crc kubenswrapper[4733]: I1204 17:44:19.101217 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:19 crc kubenswrapper[4733]: I1204 17:44:19.334299 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85b9785f75-r4tkq"] Dec 04 17:44:19 crc kubenswrapper[4733]: W1204 17:44:19.343742 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ad400c1_e98e_4f8e_b26b_9bdb007d281b.slice/crio-9e2f0c005944343c905477e34b41b8fde903be7fb42b422523451c1d0685de87 WatchSource:0}: Error finding container 9e2f0c005944343c905477e34b41b8fde903be7fb42b422523451c1d0685de87: Status 404 returned error can't find the container with id 9e2f0c005944343c905477e34b41b8fde903be7fb42b422523451c1d0685de87 Dec 04 17:44:19 crc kubenswrapper[4733]: I1204 17:44:19.376171 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm"] Dec 04 17:44:19 crc kubenswrapper[4733]: W1204 17:44:19.382359 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda512712c_0f23_47a9_8293_f66628fa3605.slice/crio-ab0b3b1f44d4778eb6b92c5cc0ac678cd64874fe5c66cb480467261c106493f4 WatchSource:0}: Error finding container ab0b3b1f44d4778eb6b92c5cc0ac678cd64874fe5c66cb480467261c106493f4: Status 404 returned error can't find the container with id ab0b3b1f44d4778eb6b92c5cc0ac678cd64874fe5c66cb480467261c106493f4 Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.009595 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" event={"ID":"a512712c-0f23-47a9-8293-f66628fa3605","Type":"ContainerStarted","Data":"1d8b2e760eae5cdd3fd4886b60aaf67756afe344d2e19fa5344cc52d1b6e9370"} Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.009960 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.009973 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" event={"ID":"a512712c-0f23-47a9-8293-f66628fa3605","Type":"ContainerStarted","Data":"ab0b3b1f44d4778eb6b92c5cc0ac678cd64874fe5c66cb480467261c106493f4"} Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.011837 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" event={"ID":"8ad400c1-e98e-4f8e-b26b-9bdb007d281b","Type":"ContainerStarted","Data":"11a13c640135ffb3a1ef76bd20166fed63945e0f11060db0e1a2e61e127c79b3"} Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.011873 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" event={"ID":"8ad400c1-e98e-4f8e-b26b-9bdb007d281b","Type":"ContainerStarted","Data":"9e2f0c005944343c905477e34b41b8fde903be7fb42b422523451c1d0685de87"} Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.012075 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.014888 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.016027 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:20 crc kubenswrapper[4733]: I1204 17:44:20.034754 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f564968d7-fngcm" podStartSLOduration=3.034522601 podStartE2EDuration="3.034522601s" podCreationTimestamp="2025-12-04 17:44:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:44:20.030595291 +0000 UTC m=+321.985956347" watchObservedRunningTime="2025-12-04 17:44:20.034522601 +0000 UTC m=+321.989883657" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.184702 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" podStartSLOduration=40.184683981 podStartE2EDuration="40.184683981s" podCreationTimestamp="2025-12-04 17:44:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:44:20.076337885 +0000 UTC m=+322.031698941" watchObservedRunningTime="2025-12-04 17:44:57.184683981 +0000 UTC m=+359.140045037" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.186016 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-85b9785f75-r4tkq"] Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.186226 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" podUID="8ad400c1-e98e-4f8e-b26b-9bdb007d281b" containerName="controller-manager" containerID="cri-o://11a13c640135ffb3a1ef76bd20166fed63945e0f11060db0e1a2e61e127c79b3" gracePeriod=30 Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.331849 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ad400c1-e98e-4f8e-b26b-9bdb007d281b" containerID="11a13c640135ffb3a1ef76bd20166fed63945e0f11060db0e1a2e61e127c79b3" exitCode=0 Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.331899 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" event={"ID":"8ad400c1-e98e-4f8e-b26b-9bdb007d281b","Type":"ContainerDied","Data":"11a13c640135ffb3a1ef76bd20166fed63945e0f11060db0e1a2e61e127c79b3"} Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.579213 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.593924 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-624lv\" (UniqueName: \"kubernetes.io/projected/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-kube-api-access-624lv\") pod \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.593984 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-serving-cert\") pod \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.594015 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-config\") pod \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.594032 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-proxy-ca-bundles\") pod \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.594056 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-client-ca\") pod \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\" (UID: \"8ad400c1-e98e-4f8e-b26b-9bdb007d281b\") " Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.594776 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-client-ca" (OuterVolumeSpecName: "client-ca") pod "8ad400c1-e98e-4f8e-b26b-9bdb007d281b" (UID: "8ad400c1-e98e-4f8e-b26b-9bdb007d281b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.594888 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.594965 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-config" (OuterVolumeSpecName: "config") pod "8ad400c1-e98e-4f8e-b26b-9bdb007d281b" (UID: "8ad400c1-e98e-4f8e-b26b-9bdb007d281b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.595365 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8ad400c1-e98e-4f8e-b26b-9bdb007d281b" (UID: "8ad400c1-e98e-4f8e-b26b-9bdb007d281b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.600287 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-kube-api-access-624lv" (OuterVolumeSpecName: "kube-api-access-624lv") pod "8ad400c1-e98e-4f8e-b26b-9bdb007d281b" (UID: "8ad400c1-e98e-4f8e-b26b-9bdb007d281b"). InnerVolumeSpecName "kube-api-access-624lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.601934 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8ad400c1-e98e-4f8e-b26b-9bdb007d281b" (UID: "8ad400c1-e98e-4f8e-b26b-9bdb007d281b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.695863 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-624lv\" (UniqueName: \"kubernetes.io/projected/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-kube-api-access-624lv\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.695897 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.695907 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:57 crc kubenswrapper[4733]: I1204 17:44:57.695916 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ad400c1-e98e-4f8e-b26b-9bdb007d281b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.344476 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.351308 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85b9785f75-r4tkq" event={"ID":"8ad400c1-e98e-4f8e-b26b-9bdb007d281b","Type":"ContainerDied","Data":"9e2f0c005944343c905477e34b41b8fde903be7fb42b422523451c1d0685de87"} Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.351355 4733 scope.go:117] "RemoveContainer" containerID="11a13c640135ffb3a1ef76bd20166fed63945e0f11060db0e1a2e61e127c79b3" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.398255 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-85b9785f75-r4tkq"] Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.404607 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-85b9785f75-r4tkq"] Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.765128 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9"] Dec 04 17:44:58 crc kubenswrapper[4733]: E1204 17:44:58.765335 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad400c1-e98e-4f8e-b26b-9bdb007d281b" containerName="controller-manager" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.765346 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad400c1-e98e-4f8e-b26b-9bdb007d281b" containerName="controller-manager" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.765441 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad400c1-e98e-4f8e-b26b-9bdb007d281b" containerName="controller-manager" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.765773 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.769288 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.771052 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.771349 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.771369 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.772236 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.772882 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9"] Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.779902 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.809023 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.811562 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bba055f-66e4-4fbc-9a37-4e8346fe9123-serving-cert\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.811624 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-config\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.811662 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-proxy-ca-bundles\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.811684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdwlj\" (UniqueName: \"kubernetes.io/projected/1bba055f-66e4-4fbc-9a37-4e8346fe9123-kube-api-access-jdwlj\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.811701 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-client-ca\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.912771 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-client-ca\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.912954 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bba055f-66e4-4fbc-9a37-4e8346fe9123-serving-cert\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.913050 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-config\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.913142 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-proxy-ca-bundles\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.913203 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdwlj\" (UniqueName: \"kubernetes.io/projected/1bba055f-66e4-4fbc-9a37-4e8346fe9123-kube-api-access-jdwlj\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.914086 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-client-ca\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.915473 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-config\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.915743 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1bba055f-66e4-4fbc-9a37-4e8346fe9123-proxy-ca-bundles\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.925574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bba055f-66e4-4fbc-9a37-4e8346fe9123-serving-cert\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:58 crc kubenswrapper[4733]: I1204 17:44:58.935522 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdwlj\" (UniqueName: \"kubernetes.io/projected/1bba055f-66e4-4fbc-9a37-4e8346fe9123-kube-api-access-jdwlj\") pod \"controller-manager-6b5b5dfbc5-hx9b9\" (UID: \"1bba055f-66e4-4fbc-9a37-4e8346fe9123\") " pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:59 crc kubenswrapper[4733]: I1204 17:44:59.096012 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:44:59 crc kubenswrapper[4733]: I1204 17:44:59.364358 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9"] Dec 04 17:44:59 crc kubenswrapper[4733]: W1204 17:44:59.370363 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bba055f_66e4_4fbc_9a37_4e8346fe9123.slice/crio-3eeacb2e2596b4c25a038d3fd4365ca1647f44efb8b31e1ec63097a71ac2cb39 WatchSource:0}: Error finding container 3eeacb2e2596b4c25a038d3fd4365ca1647f44efb8b31e1ec63097a71ac2cb39: Status 404 returned error can't find the container with id 3eeacb2e2596b4c25a038d3fd4365ca1647f44efb8b31e1ec63097a71ac2cb39 Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.178589 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g"] Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.180325 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.182073 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.190367 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g"] Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.215120 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.229902 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/535164e0-79d0-4b37-8e01-c84932cf56d6-secret-volume\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.229956 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/535164e0-79d0-4b37-8e01-c84932cf56d6-config-volume\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.230342 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/535164e0-79d0-4b37-8e01-c84932cf56d6-kube-api-access-kghvc\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.331806 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/535164e0-79d0-4b37-8e01-c84932cf56d6-secret-volume\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.331900 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/535164e0-79d0-4b37-8e01-c84932cf56d6-config-volume\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.332101 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/535164e0-79d0-4b37-8e01-c84932cf56d6-kube-api-access-kghvc\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.333477 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/535164e0-79d0-4b37-8e01-c84932cf56d6-config-volume\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.337375 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/535164e0-79d0-4b37-8e01-c84932cf56d6-secret-volume\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.346118 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad400c1-e98e-4f8e-b26b-9bdb007d281b" path="/var/lib/kubelet/pods/8ad400c1-e98e-4f8e-b26b-9bdb007d281b/volumes" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.352404 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/535164e0-79d0-4b37-8e01-c84932cf56d6-kube-api-access-kghvc\") pod \"collect-profiles-29414505-47z5g\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.356472 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" event={"ID":"1bba055f-66e4-4fbc-9a37-4e8346fe9123","Type":"ContainerStarted","Data":"20c229d460c2d2047acc406d29f392785c50d6c91615ad391a9ffb11de5919a1"} Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.356511 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" event={"ID":"1bba055f-66e4-4fbc-9a37-4e8346fe9123","Type":"ContainerStarted","Data":"3eeacb2e2596b4c25a038d3fd4365ca1647f44efb8b31e1ec63097a71ac2cb39"} Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.357042 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.361833 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.374492 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b5b5dfbc5-hx9b9" podStartSLOduration=3.3744697869999998 podStartE2EDuration="3.374469787s" podCreationTimestamp="2025-12-04 17:44:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:45:00.372954475 +0000 UTC m=+362.328315531" watchObservedRunningTime="2025-12-04 17:45:00.374469787 +0000 UTC m=+362.329830853" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.535025 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:00 crc kubenswrapper[4733]: I1204 17:45:00.984332 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g"] Dec 04 17:45:01 crc kubenswrapper[4733]: I1204 17:45:01.364284 4733 generic.go:334] "Generic (PLEG): container finished" podID="535164e0-79d0-4b37-8e01-c84932cf56d6" containerID="0ffa215e4056fa6b8382d834afca207dc6aa1683cb1451763ab5ea6daba146f8" exitCode=0 Dec 04 17:45:01 crc kubenswrapper[4733]: I1204 17:45:01.364353 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" event={"ID":"535164e0-79d0-4b37-8e01-c84932cf56d6","Type":"ContainerDied","Data":"0ffa215e4056fa6b8382d834afca207dc6aa1683cb1451763ab5ea6daba146f8"} Dec 04 17:45:01 crc kubenswrapper[4733]: I1204 17:45:01.364651 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" event={"ID":"535164e0-79d0-4b37-8e01-c84932cf56d6","Type":"ContainerStarted","Data":"bb4aaf824f3f499cd854de6a93fcd6c3c6c44bbbad458a11e1808779469526e2"} Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.658155 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.768000 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/535164e0-79d0-4b37-8e01-c84932cf56d6-config-volume\") pod \"535164e0-79d0-4b37-8e01-c84932cf56d6\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.768499 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/535164e0-79d0-4b37-8e01-c84932cf56d6-secret-volume\") pod \"535164e0-79d0-4b37-8e01-c84932cf56d6\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.768604 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/535164e0-79d0-4b37-8e01-c84932cf56d6-kube-api-access-kghvc\") pod \"535164e0-79d0-4b37-8e01-c84932cf56d6\" (UID: \"535164e0-79d0-4b37-8e01-c84932cf56d6\") " Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.768910 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/535164e0-79d0-4b37-8e01-c84932cf56d6-config-volume" (OuterVolumeSpecName: "config-volume") pod "535164e0-79d0-4b37-8e01-c84932cf56d6" (UID: "535164e0-79d0-4b37-8e01-c84932cf56d6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.775252 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535164e0-79d0-4b37-8e01-c84932cf56d6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "535164e0-79d0-4b37-8e01-c84932cf56d6" (UID: "535164e0-79d0-4b37-8e01-c84932cf56d6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.779872 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535164e0-79d0-4b37-8e01-c84932cf56d6-kube-api-access-kghvc" (OuterVolumeSpecName: "kube-api-access-kghvc") pod "535164e0-79d0-4b37-8e01-c84932cf56d6" (UID: "535164e0-79d0-4b37-8e01-c84932cf56d6"). InnerVolumeSpecName "kube-api-access-kghvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.869858 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kghvc\" (UniqueName: \"kubernetes.io/projected/535164e0-79d0-4b37-8e01-c84932cf56d6-kube-api-access-kghvc\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.869911 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/535164e0-79d0-4b37-8e01-c84932cf56d6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:02 crc kubenswrapper[4733]: I1204 17:45:02.869929 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/535164e0-79d0-4b37-8e01-c84932cf56d6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:03 crc kubenswrapper[4733]: I1204 17:45:03.378058 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" event={"ID":"535164e0-79d0-4b37-8e01-c84932cf56d6","Type":"ContainerDied","Data":"bb4aaf824f3f499cd854de6a93fcd6c3c6c44bbbad458a11e1808779469526e2"} Dec 04 17:45:03 crc kubenswrapper[4733]: I1204 17:45:03.378106 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb4aaf824f3f499cd854de6a93fcd6c3c6c44bbbad458a11e1808779469526e2" Dec 04 17:45:03 crc kubenswrapper[4733]: I1204 17:45:03.378190 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.056859 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-llz25"] Dec 04 17:45:08 crc kubenswrapper[4733]: E1204 17:45:08.057092 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535164e0-79d0-4b37-8e01-c84932cf56d6" containerName="collect-profiles" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.057106 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="535164e0-79d0-4b37-8e01-c84932cf56d6" containerName="collect-profiles" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.057230 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="535164e0-79d0-4b37-8e01-c84932cf56d6" containerName="collect-profiles" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.057710 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.089663 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-llz25"] Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.240619 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-bound-sa-token\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.240672 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b8bc3117-39da-48f3-ba9c-ecc880df4885-registry-certificates\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.240700 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s9kw\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-kube-api-access-4s9kw\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.240736 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.240781 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b8bc3117-39da-48f3-ba9c-ecc880df4885-ca-trust-extracted\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.240846 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b8bc3117-39da-48f3-ba9c-ecc880df4885-installation-pull-secrets\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.240902 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-registry-tls\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.240934 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b8bc3117-39da-48f3-ba9c-ecc880df4885-trusted-ca\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.278041 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.341515 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-registry-tls\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.341554 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b8bc3117-39da-48f3-ba9c-ecc880df4885-trusted-ca\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.341594 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-bound-sa-token\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.341613 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b8bc3117-39da-48f3-ba9c-ecc880df4885-registry-certificates\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.341632 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s9kw\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-kube-api-access-4s9kw\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.341652 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b8bc3117-39da-48f3-ba9c-ecc880df4885-ca-trust-extracted\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.341684 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b8bc3117-39da-48f3-ba9c-ecc880df4885-installation-pull-secrets\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.342633 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b8bc3117-39da-48f3-ba9c-ecc880df4885-ca-trust-extracted\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.343885 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b8bc3117-39da-48f3-ba9c-ecc880df4885-registry-certificates\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.344013 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b8bc3117-39da-48f3-ba9c-ecc880df4885-trusted-ca\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.348098 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b8bc3117-39da-48f3-ba9c-ecc880df4885-installation-pull-secrets\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.352360 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-registry-tls\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.365349 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s9kw\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-kube-api-access-4s9kw\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.374199 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b8bc3117-39da-48f3-ba9c-ecc880df4885-bound-sa-token\") pod \"image-registry-66df7c8f76-llz25\" (UID: \"b8bc3117-39da-48f3-ba9c-ecc880df4885\") " pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.382395 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:08 crc kubenswrapper[4733]: I1204 17:45:08.896537 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-llz25"] Dec 04 17:45:09 crc kubenswrapper[4733]: I1204 17:45:09.417353 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-llz25" event={"ID":"b8bc3117-39da-48f3-ba9c-ecc880df4885","Type":"ContainerStarted","Data":"ca71c9f894d9849146dc77a195e748c60fcd89d235863f4eb538ed9793b60fdf"} Dec 04 17:45:09 crc kubenswrapper[4733]: I1204 17:45:09.417754 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:09 crc kubenswrapper[4733]: I1204 17:45:09.417768 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-llz25" event={"ID":"b8bc3117-39da-48f3-ba9c-ecc880df4885","Type":"ContainerStarted","Data":"da00555b9024aff9f1401b363d4c9d69f59188490612d5aca75e7f0364df7008"} Dec 04 17:45:09 crc kubenswrapper[4733]: I1204 17:45:09.438594 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-llz25" podStartSLOduration=1.438566709 podStartE2EDuration="1.438566709s" podCreationTimestamp="2025-12-04 17:45:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:45:09.435525935 +0000 UTC m=+371.390887021" watchObservedRunningTime="2025-12-04 17:45:09.438566709 +0000 UTC m=+371.393927795" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.321770 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v2fw4"] Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.322444 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v2fw4" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerName="registry-server" containerID="cri-o://8d131100c38d0de1ba07ae88f82c29e47f56606933dbf747d75ad0158efe3f11" gracePeriod=30 Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.331930 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-29x4q"] Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.332171 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-29x4q" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="registry-server" containerID="cri-o://aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd" gracePeriod=30 Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.343973 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgk9"] Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.344393 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerName="marketplace-operator" containerID="cri-o://2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526" gracePeriod=30 Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.350827 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgmp"] Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.351091 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xkgmp" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerName="registry-server" containerID="cri-o://78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d" gracePeriod=30 Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.355488 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gpbpj"] Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.355917 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gpbpj" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" containerName="registry-server" containerID="cri-o://2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8" gracePeriod=30 Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.361692 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-545tk"] Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.365188 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.369872 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-545tk"] Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.446982 4733 generic.go:334] "Generic (PLEG): container finished" podID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerID="8d131100c38d0de1ba07ae88f82c29e47f56606933dbf747d75ad0158efe3f11" exitCode=0 Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.447042 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2fw4" event={"ID":"26041ca9-71b9-4deb-a4e0-8192e7bac916","Type":"ContainerDied","Data":"8d131100c38d0de1ba07ae88f82c29e47f56606933dbf747d75ad0158efe3f11"} Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.519674 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjqvd\" (UniqueName: \"kubernetes.io/projected/17e52f55-0856-43c8-912b-6594e33aaebc-kube-api-access-gjqvd\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.519749 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17e52f55-0856-43c8-912b-6594e33aaebc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.519832 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17e52f55-0856-43c8-912b-6594e33aaebc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.621311 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17e52f55-0856-43c8-912b-6594e33aaebc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.621349 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17e52f55-0856-43c8-912b-6594e33aaebc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.621397 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjqvd\" (UniqueName: \"kubernetes.io/projected/17e52f55-0856-43c8-912b-6594e33aaebc-kube-api-access-gjqvd\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.622826 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17e52f55-0856-43c8-912b-6594e33aaebc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.631429 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17e52f55-0856-43c8-912b-6594e33aaebc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.640507 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjqvd\" (UniqueName: \"kubernetes.io/projected/17e52f55-0856-43c8-912b-6594e33aaebc-kube-api-access-gjqvd\") pod \"marketplace-operator-79b997595-545tk\" (UID: \"17e52f55-0856-43c8-912b-6594e33aaebc\") " pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.691123 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:13 crc kubenswrapper[4733]: I1204 17:45:13.878136 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.021766 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.027008 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.035840 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-trusted-ca\") pod \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.035933 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkql8\" (UniqueName: \"kubernetes.io/projected/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-kube-api-access-fkql8\") pod \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.035988 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-operator-metrics\") pod \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\" (UID: \"7caf3411-fca7-4025-b9b6-eb3b4a43ab81\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.041464 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7caf3411-fca7-4025-b9b6-eb3b4a43ab81" (UID: "7caf3411-fca7-4025-b9b6-eb3b4a43ab81"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.044869 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-kube-api-access-fkql8" (OuterVolumeSpecName: "kube-api-access-fkql8") pod "7caf3411-fca7-4025-b9b6-eb3b4a43ab81" (UID: "7caf3411-fca7-4025-b9b6-eb3b4a43ab81"). InnerVolumeSpecName "kube-api-access-fkql8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.045245 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7caf3411-fca7-4025-b9b6-eb3b4a43ab81" (UID: "7caf3411-fca7-4025-b9b6-eb3b4a43ab81"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.045548 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.074410 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137245 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-utilities\") pod \"26041ca9-71b9-4deb-a4e0-8192e7bac916\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137319 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzdk4\" (UniqueName: \"kubernetes.io/projected/26041ca9-71b9-4deb-a4e0-8192e7bac916-kube-api-access-jzdk4\") pod \"26041ca9-71b9-4deb-a4e0-8192e7bac916\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137349 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnlzg\" (UniqueName: \"kubernetes.io/projected/d8eae688-012f-4c73-bd36-828cb9350458-kube-api-access-dnlzg\") pod \"d8eae688-012f-4c73-bd36-828cb9350458\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137379 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-catalog-content\") pod \"d8eae688-012f-4c73-bd36-828cb9350458\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137394 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58xwk\" (UniqueName: \"kubernetes.io/projected/1e13e0b3-c69f-4494-9e11-480d9ced64b8-kube-api-access-58xwk\") pod \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137412 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-utilities\") pod \"d8eae688-012f-4c73-bd36-828cb9350458\" (UID: \"d8eae688-012f-4c73-bd36-828cb9350458\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137444 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-catalog-content\") pod \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137458 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-catalog-content\") pod \"26041ca9-71b9-4deb-a4e0-8192e7bac916\" (UID: \"26041ca9-71b9-4deb-a4e0-8192e7bac916\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137484 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-utilities\") pod \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\" (UID: \"1e13e0b3-c69f-4494-9e11-480d9ced64b8\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137671 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkql8\" (UniqueName: \"kubernetes.io/projected/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-kube-api-access-fkql8\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137682 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.137691 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7caf3411-fca7-4025-b9b6-eb3b4a43ab81-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.138424 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-utilities" (OuterVolumeSpecName: "utilities") pod "26041ca9-71b9-4deb-a4e0-8192e7bac916" (UID: "26041ca9-71b9-4deb-a4e0-8192e7bac916"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.140560 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26041ca9-71b9-4deb-a4e0-8192e7bac916-kube-api-access-jzdk4" (OuterVolumeSpecName: "kube-api-access-jzdk4") pod "26041ca9-71b9-4deb-a4e0-8192e7bac916" (UID: "26041ca9-71b9-4deb-a4e0-8192e7bac916"). InnerVolumeSpecName "kube-api-access-jzdk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.141256 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-utilities" (OuterVolumeSpecName: "utilities") pod "1e13e0b3-c69f-4494-9e11-480d9ced64b8" (UID: "1e13e0b3-c69f-4494-9e11-480d9ced64b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.142170 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8eae688-012f-4c73-bd36-828cb9350458-kube-api-access-dnlzg" (OuterVolumeSpecName: "kube-api-access-dnlzg") pod "d8eae688-012f-4c73-bd36-828cb9350458" (UID: "d8eae688-012f-4c73-bd36-828cb9350458"). InnerVolumeSpecName "kube-api-access-dnlzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.142746 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e13e0b3-c69f-4494-9e11-480d9ced64b8-kube-api-access-58xwk" (OuterVolumeSpecName: "kube-api-access-58xwk") pod "1e13e0b3-c69f-4494-9e11-480d9ced64b8" (UID: "1e13e0b3-c69f-4494-9e11-480d9ced64b8"). InnerVolumeSpecName "kube-api-access-58xwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.143379 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-utilities" (OuterVolumeSpecName: "utilities") pod "d8eae688-012f-4c73-bd36-828cb9350458" (UID: "d8eae688-012f-4c73-bd36-828cb9350458"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.159393 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e13e0b3-c69f-4494-9e11-480d9ced64b8" (UID: "1e13e0b3-c69f-4494-9e11-480d9ced64b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.188051 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26041ca9-71b9-4deb-a4e0-8192e7bac916" (UID: "26041ca9-71b9-4deb-a4e0-8192e7bac916"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.200193 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8eae688-012f-4c73-bd36-828cb9350458" (UID: "d8eae688-012f-4c73-bd36-828cb9350458"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.209181 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-545tk"] Dec 04 17:45:14 crc kubenswrapper[4733]: W1204 17:45:14.217022 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17e52f55_0856_43c8_912b_6594e33aaebc.slice/crio-69f16b719e08173028779b217d7afd0ca23bc1c4fa96041f1c21a82efdd18676 WatchSource:0}: Error finding container 69f16b719e08173028779b217d7afd0ca23bc1c4fa96041f1c21a82efdd18676: Status 404 returned error can't find the container with id 69f16b719e08173028779b217d7afd0ca23bc1c4fa96041f1c21a82efdd18676 Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.238857 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5mqp\" (UniqueName: \"kubernetes.io/projected/c19dfb55-c152-406a-b35e-35498a8b7690-kube-api-access-l5mqp\") pod \"c19dfb55-c152-406a-b35e-35498a8b7690\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.238905 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-utilities\") pod \"c19dfb55-c152-406a-b35e-35498a8b7690\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.238954 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-catalog-content\") pod \"c19dfb55-c152-406a-b35e-35498a8b7690\" (UID: \"c19dfb55-c152-406a-b35e-35498a8b7690\") " Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239227 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzdk4\" (UniqueName: \"kubernetes.io/projected/26041ca9-71b9-4deb-a4e0-8192e7bac916-kube-api-access-jzdk4\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239251 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnlzg\" (UniqueName: \"kubernetes.io/projected/d8eae688-012f-4c73-bd36-828cb9350458-kube-api-access-dnlzg\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239264 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239276 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58xwk\" (UniqueName: \"kubernetes.io/projected/1e13e0b3-c69f-4494-9e11-480d9ced64b8-kube-api-access-58xwk\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239288 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8eae688-012f-4c73-bd36-828cb9350458-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239299 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239309 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239321 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e13e0b3-c69f-4494-9e11-480d9ced64b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239332 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26041ca9-71b9-4deb-a4e0-8192e7bac916-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.239951 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-utilities" (OuterVolumeSpecName: "utilities") pod "c19dfb55-c152-406a-b35e-35498a8b7690" (UID: "c19dfb55-c152-406a-b35e-35498a8b7690"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.247104 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19dfb55-c152-406a-b35e-35498a8b7690-kube-api-access-l5mqp" (OuterVolumeSpecName: "kube-api-access-l5mqp") pod "c19dfb55-c152-406a-b35e-35498a8b7690" (UID: "c19dfb55-c152-406a-b35e-35498a8b7690"). InnerVolumeSpecName "kube-api-access-l5mqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.340471 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5mqp\" (UniqueName: \"kubernetes.io/projected/c19dfb55-c152-406a-b35e-35498a8b7690-kube-api-access-l5mqp\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.340488 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.347063 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c19dfb55-c152-406a-b35e-35498a8b7690" (UID: "c19dfb55-c152-406a-b35e-35498a8b7690"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.441858 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19dfb55-c152-406a-b35e-35498a8b7690-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.460342 4733 generic.go:334] "Generic (PLEG): container finished" podID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerID="78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d" exitCode=0 Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.460394 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgmp" event={"ID":"1e13e0b3-c69f-4494-9e11-480d9ced64b8","Type":"ContainerDied","Data":"78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.460422 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgmp" event={"ID":"1e13e0b3-c69f-4494-9e11-480d9ced64b8","Type":"ContainerDied","Data":"300a0b4d47c5054adf89b67e3c7aef638f4c8ce347a62726e7af2fac333fb7e1"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.460439 4733 scope.go:117] "RemoveContainer" containerID="78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.460542 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkgmp" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.466764 4733 generic.go:334] "Generic (PLEG): container finished" podID="c19dfb55-c152-406a-b35e-35498a8b7690" containerID="2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8" exitCode=0 Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.466904 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gpbpj" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.466921 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpbpj" event={"ID":"c19dfb55-c152-406a-b35e-35498a8b7690","Type":"ContainerDied","Data":"2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.466977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gpbpj" event={"ID":"c19dfb55-c152-406a-b35e-35498a8b7690","Type":"ContainerDied","Data":"797f5b525049d270e0df6192f142c6b2e334f7717b043754778c5a59a38f70fb"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.471150 4733 generic.go:334] "Generic (PLEG): container finished" podID="d8eae688-012f-4c73-bd36-828cb9350458" containerID="aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd" exitCode=0 Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.471214 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29x4q" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.471243 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29x4q" event={"ID":"d8eae688-012f-4c73-bd36-828cb9350458","Type":"ContainerDied","Data":"aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.471276 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29x4q" event={"ID":"d8eae688-012f-4c73-bd36-828cb9350458","Type":"ContainerDied","Data":"ae65f4efa5533f262462cc5107e95b5d5ba7355329c8f27b1e8412232a23264f"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.475467 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2fw4" event={"ID":"26041ca9-71b9-4deb-a4e0-8192e7bac916","Type":"ContainerDied","Data":"a8888e8306681261b51b11705bd15cf618ef20b2e7f4da8a488642504c2d1427"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.475713 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2fw4" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.478222 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-545tk" event={"ID":"17e52f55-0856-43c8-912b-6594e33aaebc","Type":"ContainerStarted","Data":"f2981d65f8b564d0333ebb39ca7242d592a2e5922a5e05748976ec3d36a3a2e5"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.478313 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-545tk" event={"ID":"17e52f55-0856-43c8-912b-6594e33aaebc","Type":"ContainerStarted","Data":"69f16b719e08173028779b217d7afd0ca23bc1c4fa96041f1c21a82efdd18676"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.478504 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.479735 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-545tk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" start-of-body= Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.479860 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-545tk" podUID="17e52f55-0856-43c8-912b-6594e33aaebc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.481532 4733 generic.go:334] "Generic (PLEG): container finished" podID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerID="2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526" exitCode=0 Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.481617 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.481590 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" event={"ID":"7caf3411-fca7-4025-b9b6-eb3b4a43ab81","Type":"ContainerDied","Data":"2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.482583 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgk9" event={"ID":"7caf3411-fca7-4025-b9b6-eb3b4a43ab81","Type":"ContainerDied","Data":"8aeff507b88a4b15055d0b3b7717627cc13d40ef5a348bdc3eb9bec4f5534cd6"} Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.518593 4733 scope.go:117] "RemoveContainer" containerID="bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.523296 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-545tk" podStartSLOduration=1.5232797 podStartE2EDuration="1.5232797s" podCreationTimestamp="2025-12-04 17:45:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:45:14.519040592 +0000 UTC m=+376.474401638" watchObservedRunningTime="2025-12-04 17:45:14.5232797 +0000 UTC m=+376.478640756" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.545913 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-29x4q"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.561406 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-29x4q"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.573320 4733 scope.go:117] "RemoveContainer" containerID="2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.573462 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v2fw4"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.593591 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v2fw4"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.598062 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgmp"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.603653 4733 scope.go:117] "RemoveContainer" containerID="78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.604218 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d\": container with ID starting with 78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d not found: ID does not exist" containerID="78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.604258 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d"} err="failed to get container status \"78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d\": rpc error: code = NotFound desc = could not find container \"78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d\": container with ID starting with 78f1a5e753bf3da62d50a6d5f6b8cb746fec2454dc303c90a772592485705f0d not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.604287 4733 scope.go:117] "RemoveContainer" containerID="bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.604735 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad\": container with ID starting with bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad not found: ID does not exist" containerID="bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.604777 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad"} err="failed to get container status \"bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad\": rpc error: code = NotFound desc = could not find container \"bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad\": container with ID starting with bbf56f7ab960d91a56341bbd0929d8303bcca0c66ef918d9ee42a4959a620fad not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.604859 4733 scope.go:117] "RemoveContainer" containerID="2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.605131 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb\": container with ID starting with 2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb not found: ID does not exist" containerID="2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.605166 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb"} err="failed to get container status \"2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb\": rpc error: code = NotFound desc = could not find container \"2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb\": container with ID starting with 2e2998fa7eefb0ee82440d725017b2f975ae3c048dfdc481fde9a1813be290cb not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.605188 4733 scope.go:117] "RemoveContainer" containerID="2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.605543 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgmp"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.609942 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gpbpj"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.614933 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gpbpj"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.619686 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgk9"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.622438 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgk9"] Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.626354 4733 scope.go:117] "RemoveContainer" containerID="8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.642835 4733 scope.go:117] "RemoveContainer" containerID="0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.655184 4733 scope.go:117] "RemoveContainer" containerID="2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.656359 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8\": container with ID starting with 2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8 not found: ID does not exist" containerID="2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.656396 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8"} err="failed to get container status \"2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8\": rpc error: code = NotFound desc = could not find container \"2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8\": container with ID starting with 2d0e1ef5335d79616acf5c83f7f23eeb019aa44a0dca03435e9be983610791b8 not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.656427 4733 scope.go:117] "RemoveContainer" containerID="8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.656949 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44\": container with ID starting with 8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44 not found: ID does not exist" containerID="8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.657001 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44"} err="failed to get container status \"8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44\": rpc error: code = NotFound desc = could not find container \"8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44\": container with ID starting with 8495c338dd9c04d5322a2082c04040dc2ccf8e1bf240d0df3b3a83a4b7b7ac44 not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.657030 4733 scope.go:117] "RemoveContainer" containerID="0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.657925 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6\": container with ID starting with 0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6 not found: ID does not exist" containerID="0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.657951 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6"} err="failed to get container status \"0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6\": rpc error: code = NotFound desc = could not find container \"0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6\": container with ID starting with 0b67a814d0c16625cbc27a5ed2d0a84a8ef35f105684e4a8216357c3b53ed4e6 not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.657966 4733 scope.go:117] "RemoveContainer" containerID="aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.673119 4733 scope.go:117] "RemoveContainer" containerID="45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.689380 4733 scope.go:117] "RemoveContainer" containerID="de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.701841 4733 scope.go:117] "RemoveContainer" containerID="aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.702170 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd\": container with ID starting with aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd not found: ID does not exist" containerID="aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.702197 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd"} err="failed to get container status \"aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd\": rpc error: code = NotFound desc = could not find container \"aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd\": container with ID starting with aa5badd7e9735aa4d7c4876410c2575087b2d2fbe284832ec6aa1d3022c9c3dd not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.702223 4733 scope.go:117] "RemoveContainer" containerID="45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.702392 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938\": container with ID starting with 45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938 not found: ID does not exist" containerID="45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.702409 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938"} err="failed to get container status \"45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938\": rpc error: code = NotFound desc = could not find container \"45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938\": container with ID starting with 45621af8a70edf705d1cfb1e02c0175ae1da714a58241942e7ebb65b4127e938 not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.702421 4733 scope.go:117] "RemoveContainer" containerID="de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.702597 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e\": container with ID starting with de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e not found: ID does not exist" containerID="de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.702614 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e"} err="failed to get container status \"de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e\": rpc error: code = NotFound desc = could not find container \"de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e\": container with ID starting with de6bf78b98e86cded00c1fc0f9c56b2a80d21b7b64bceca3900948c98753ec9e not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.702628 4733 scope.go:117] "RemoveContainer" containerID="8d131100c38d0de1ba07ae88f82c29e47f56606933dbf747d75ad0158efe3f11" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.715578 4733 scope.go:117] "RemoveContainer" containerID="7e5b33371b5190df651ec97ac9d4cc20ac72a8067931ad0c515c343240748cc0" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.728210 4733 scope.go:117] "RemoveContainer" containerID="6861c904bf894b593b0c32dfe97a422b55ff9699ac08100ecc408df623274cc9" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.739206 4733 scope.go:117] "RemoveContainer" containerID="2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.751183 4733 scope.go:117] "RemoveContainer" containerID="5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.765356 4733 scope.go:117] "RemoveContainer" containerID="2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.765678 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526\": container with ID starting with 2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526 not found: ID does not exist" containerID="2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.765707 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526"} err="failed to get container status \"2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526\": rpc error: code = NotFound desc = could not find container \"2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526\": container with ID starting with 2fb26348141d5f21a6fec797b842660483695929e7412fadaa0b088745680526 not found: ID does not exist" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.765728 4733 scope.go:117] "RemoveContainer" containerID="5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3" Dec 04 17:45:14 crc kubenswrapper[4733]: E1204 17:45:14.766025 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3\": container with ID starting with 5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3 not found: ID does not exist" containerID="5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3" Dec 04 17:45:14 crc kubenswrapper[4733]: I1204 17:45:14.766067 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3"} err="failed to get container status \"5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3\": rpc error: code = NotFound desc = could not find container \"5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3\": container with ID starting with 5dd8c79280ada90d0652b87e0f6e531bab04a863b2ab72a623af71a508376bd3 not found: ID does not exist" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.362412 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.362498 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.505564 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-545tk" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.936765 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-npgrs"] Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.936994 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerName="extract-utilities" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937008 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerName="extract-utilities" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937020 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937028 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937038 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="extract-content" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937048 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="extract-content" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937061 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerName="extract-content" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937068 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerName="extract-content" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937078 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937086 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937097 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="extract-utilities" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937105 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="extract-utilities" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937119 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937129 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937140 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" containerName="extract-content" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937148 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" containerName="extract-content" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937161 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerName="marketplace-operator" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937168 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerName="marketplace-operator" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937180 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerName="extract-utilities" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937187 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerName="extract-utilities" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937197 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerName="marketplace-operator" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937205 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerName="marketplace-operator" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937215 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" containerName="extract-utilities" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937222 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" containerName="extract-utilities" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937233 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937240 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: E1204 17:45:15.937251 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerName="extract-content" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937259 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerName="extract-content" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937358 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937375 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937385 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8eae688-012f-4c73-bd36-828cb9350458" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937397 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerName="marketplace-operator" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937407 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" containerName="registry-server" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.937416 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" containerName="marketplace-operator" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.938711 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.941318 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 17:45:15 crc kubenswrapper[4733]: I1204 17:45:15.949698 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-npgrs"] Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.066845 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46w65\" (UniqueName: \"kubernetes.io/projected/f455a245-3a28-48c0-b83a-5b5b56c897f6-kube-api-access-46w65\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.066913 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-utilities\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.066959 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-catalog-content\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.168027 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-catalog-content\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.168198 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46w65\" (UniqueName: \"kubernetes.io/projected/f455a245-3a28-48c0-b83a-5b5b56c897f6-kube-api-access-46w65\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.168238 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-utilities\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.169055 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-utilities\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.169277 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-catalog-content\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.196719 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46w65\" (UniqueName: \"kubernetes.io/projected/f455a245-3a28-48c0-b83a-5b5b56c897f6-kube-api-access-46w65\") pod \"redhat-operators-npgrs\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.259067 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.341924 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e13e0b3-c69f-4494-9e11-480d9ced64b8" path="/var/lib/kubelet/pods/1e13e0b3-c69f-4494-9e11-480d9ced64b8/volumes" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.342539 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26041ca9-71b9-4deb-a4e0-8192e7bac916" path="/var/lib/kubelet/pods/26041ca9-71b9-4deb-a4e0-8192e7bac916/volumes" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.343151 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7caf3411-fca7-4025-b9b6-eb3b4a43ab81" path="/var/lib/kubelet/pods/7caf3411-fca7-4025-b9b6-eb3b4a43ab81/volumes" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.344083 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c19dfb55-c152-406a-b35e-35498a8b7690" path="/var/lib/kubelet/pods/c19dfb55-c152-406a-b35e-35498a8b7690/volumes" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.344627 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8eae688-012f-4c73-bd36-828cb9350458" path="/var/lib/kubelet/pods/d8eae688-012f-4c73-bd36-828cb9350458/volumes" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.691093 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-npgrs"] Dec 04 17:45:16 crc kubenswrapper[4733]: W1204 17:45:16.704450 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice/crio-783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4 WatchSource:0}: Error finding container 783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4: Status 404 returned error can't find the container with id 783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4 Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.938068 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-khxrh"] Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.940087 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.942706 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 17:45:16 crc kubenswrapper[4733]: I1204 17:45:16.944164 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-khxrh"] Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.082823 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-catalog-content\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.082899 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vbxm\" (UniqueName: \"kubernetes.io/projected/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-kube-api-access-5vbxm\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.083005 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-utilities\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.184286 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-utilities\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.184369 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-catalog-content\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.184387 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vbxm\" (UniqueName: \"kubernetes.io/projected/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-kube-api-access-5vbxm\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.184970 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-catalog-content\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.185349 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-utilities\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.219517 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vbxm\" (UniqueName: \"kubernetes.io/projected/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-kube-api-access-5vbxm\") pod \"certified-operators-khxrh\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.273063 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.510467 4733 generic.go:334] "Generic (PLEG): container finished" podID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerID="bd83ef32350485919bf0a69613a5a0017daaf6be41e8e9d03557fece45e9861f" exitCode=0 Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.510603 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npgrs" event={"ID":"f455a245-3a28-48c0-b83a-5b5b56c897f6","Type":"ContainerDied","Data":"bd83ef32350485919bf0a69613a5a0017daaf6be41e8e9d03557fece45e9861f"} Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.510726 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npgrs" event={"ID":"f455a245-3a28-48c0-b83a-5b5b56c897f6","Type":"ContainerStarted","Data":"783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4"} Dec 04 17:45:17 crc kubenswrapper[4733]: I1204 17:45:17.728218 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-khxrh"] Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.355680 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cj6ms"] Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.370654 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cj6ms"] Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.370952 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.377722 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.501188 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-catalog-content\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.501346 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvdjm\" (UniqueName: \"kubernetes.io/projected/72e0892e-4912-407e-b6ee-438c98fb6d8a-kube-api-access-mvdjm\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.501426 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-utilities\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.519494 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerID="ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da" exitCode=0 Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.519566 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khxrh" event={"ID":"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d","Type":"ContainerDied","Data":"ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da"} Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.519642 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khxrh" event={"ID":"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d","Type":"ContainerStarted","Data":"5b870eaba2113352e722055c286690fc2345480266ab17b2f115cab9bf43e7a8"} Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.522102 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npgrs" event={"ID":"f455a245-3a28-48c0-b83a-5b5b56c897f6","Type":"ContainerStarted","Data":"2e7f897c5b3b9347af79d9cea190c633bca6e27299b06ceab8259937b570a9f4"} Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.602262 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvdjm\" (UniqueName: \"kubernetes.io/projected/72e0892e-4912-407e-b6ee-438c98fb6d8a-kube-api-access-mvdjm\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.602340 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-utilities\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.602378 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-catalog-content\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.603816 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-catalog-content\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.616115 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-utilities\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.635937 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvdjm\" (UniqueName: \"kubernetes.io/projected/72e0892e-4912-407e-b6ee-438c98fb6d8a-kube-api-access-mvdjm\") pod \"community-operators-cj6ms\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:18 crc kubenswrapper[4733]: I1204 17:45:18.701424 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.114164 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cj6ms"] Dec 04 17:45:19 crc kubenswrapper[4733]: W1204 17:45:19.138256 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72e0892e_4912_407e_b6ee_438c98fb6d8a.slice/crio-3a541950764629e12490f6bda2fb6e0c7103e86a63871431327713af913cd60f WatchSource:0}: Error finding container 3a541950764629e12490f6bda2fb6e0c7103e86a63871431327713af913cd60f: Status 404 returned error can't find the container with id 3a541950764629e12490f6bda2fb6e0c7103e86a63871431327713af913cd60f Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.354133 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gb2qr"] Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.356814 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.359388 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.362841 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gb2qr"] Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.515142 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d40421-3817-47d8-85c1-aeb819796b70-catalog-content\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.515188 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d40421-3817-47d8-85c1-aeb819796b70-utilities\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.515229 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmldh\" (UniqueName: \"kubernetes.io/projected/d1d40421-3817-47d8-85c1-aeb819796b70-kube-api-access-hmldh\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.529271 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khxrh" event={"ID":"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d","Type":"ContainerStarted","Data":"8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59"} Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.532031 4733 generic.go:334] "Generic (PLEG): container finished" podID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerID="2e7f897c5b3b9347af79d9cea190c633bca6e27299b06ceab8259937b570a9f4" exitCode=0 Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.532208 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npgrs" event={"ID":"f455a245-3a28-48c0-b83a-5b5b56c897f6","Type":"ContainerDied","Data":"2e7f897c5b3b9347af79d9cea190c633bca6e27299b06ceab8259937b570a9f4"} Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.535973 4733 generic.go:334] "Generic (PLEG): container finished" podID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerID="e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692" exitCode=0 Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.536026 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cj6ms" event={"ID":"72e0892e-4912-407e-b6ee-438c98fb6d8a","Type":"ContainerDied","Data":"e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692"} Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.536057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cj6ms" event={"ID":"72e0892e-4912-407e-b6ee-438c98fb6d8a","Type":"ContainerStarted","Data":"3a541950764629e12490f6bda2fb6e0c7103e86a63871431327713af913cd60f"} Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.616552 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d40421-3817-47d8-85c1-aeb819796b70-utilities\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.617035 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmldh\" (UniqueName: \"kubernetes.io/projected/d1d40421-3817-47d8-85c1-aeb819796b70-kube-api-access-hmldh\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.617228 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d40421-3817-47d8-85c1-aeb819796b70-utilities\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.617661 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d40421-3817-47d8-85c1-aeb819796b70-catalog-content\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.618119 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d40421-3817-47d8-85c1-aeb819796b70-catalog-content\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.649575 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmldh\" (UniqueName: \"kubernetes.io/projected/d1d40421-3817-47d8-85c1-aeb819796b70-kube-api-access-hmldh\") pod \"redhat-marketplace-gb2qr\" (UID: \"d1d40421-3817-47d8-85c1-aeb819796b70\") " pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:19 crc kubenswrapper[4733]: I1204 17:45:19.693462 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.115674 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gb2qr"] Dec 04 17:45:20 crc kubenswrapper[4733]: W1204 17:45:20.126009 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1d40421_3817_47d8_85c1_aeb819796b70.slice/crio-6fa1a4fd901c3fafba16f7275055d77ae903e4d57d7fdb56e992cb67f39bf9e3 WatchSource:0}: Error finding container 6fa1a4fd901c3fafba16f7275055d77ae903e4d57d7fdb56e992cb67f39bf9e3: Status 404 returned error can't find the container with id 6fa1a4fd901c3fafba16f7275055d77ae903e4d57d7fdb56e992cb67f39bf9e3 Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.542880 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npgrs" event={"ID":"f455a245-3a28-48c0-b83a-5b5b56c897f6","Type":"ContainerStarted","Data":"9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c"} Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.545200 4733 generic.go:334] "Generic (PLEG): container finished" podID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerID="d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee" exitCode=0 Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.545286 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cj6ms" event={"ID":"72e0892e-4912-407e-b6ee-438c98fb6d8a","Type":"ContainerDied","Data":"d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee"} Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.547494 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerID="8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59" exitCode=0 Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.547571 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khxrh" event={"ID":"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d","Type":"ContainerDied","Data":"8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59"} Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.550751 4733 generic.go:334] "Generic (PLEG): container finished" podID="d1d40421-3817-47d8-85c1-aeb819796b70" containerID="e91ff6345b22dbd38cb3270a5ecfd60f3295b8bdffa4d221b7e27cc0bcb6711b" exitCode=0 Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.550809 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gb2qr" event={"ID":"d1d40421-3817-47d8-85c1-aeb819796b70","Type":"ContainerDied","Data":"e91ff6345b22dbd38cb3270a5ecfd60f3295b8bdffa4d221b7e27cc0bcb6711b"} Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.550841 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gb2qr" event={"ID":"d1d40421-3817-47d8-85c1-aeb819796b70","Type":"ContainerStarted","Data":"6fa1a4fd901c3fafba16f7275055d77ae903e4d57d7fdb56e992cb67f39bf9e3"} Dec 04 17:45:20 crc kubenswrapper[4733]: I1204 17:45:20.564751 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-npgrs" podStartSLOduration=3.120849044 podStartE2EDuration="5.564728523s" podCreationTimestamp="2025-12-04 17:45:15 +0000 UTC" firstStartedPulling="2025-12-04 17:45:17.512581315 +0000 UTC m=+379.467942351" lastFinishedPulling="2025-12-04 17:45:19.956460784 +0000 UTC m=+381.911821830" observedRunningTime="2025-12-04 17:45:20.558317186 +0000 UTC m=+382.513678272" watchObservedRunningTime="2025-12-04 17:45:20.564728523 +0000 UTC m=+382.520089599" Dec 04 17:45:21 crc kubenswrapper[4733]: I1204 17:45:21.557597 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khxrh" event={"ID":"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d","Type":"ContainerStarted","Data":"169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b"} Dec 04 17:45:21 crc kubenswrapper[4733]: I1204 17:45:21.559505 4733 generic.go:334] "Generic (PLEG): container finished" podID="d1d40421-3817-47d8-85c1-aeb819796b70" containerID="fcbd8b7bde465af92de028c49aca079b0a90a546976e41a7cacf883f00a9c169" exitCode=0 Dec 04 17:45:21 crc kubenswrapper[4733]: I1204 17:45:21.559585 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gb2qr" event={"ID":"d1d40421-3817-47d8-85c1-aeb819796b70","Type":"ContainerDied","Data":"fcbd8b7bde465af92de028c49aca079b0a90a546976e41a7cacf883f00a9c169"} Dec 04 17:45:21 crc kubenswrapper[4733]: I1204 17:45:21.563575 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cj6ms" event={"ID":"72e0892e-4912-407e-b6ee-438c98fb6d8a","Type":"ContainerStarted","Data":"0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2"} Dec 04 17:45:21 crc kubenswrapper[4733]: I1204 17:45:21.574161 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-khxrh" podStartSLOduration=3.168316226 podStartE2EDuration="5.574143558s" podCreationTimestamp="2025-12-04 17:45:16 +0000 UTC" firstStartedPulling="2025-12-04 17:45:18.521354063 +0000 UTC m=+380.476715109" lastFinishedPulling="2025-12-04 17:45:20.927181365 +0000 UTC m=+382.882542441" observedRunningTime="2025-12-04 17:45:21.573887771 +0000 UTC m=+383.529248837" watchObservedRunningTime="2025-12-04 17:45:21.574143558 +0000 UTC m=+383.529504614" Dec 04 17:45:21 crc kubenswrapper[4733]: I1204 17:45:21.613740 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cj6ms" podStartSLOduration=2.1670684749999998 podStartE2EDuration="3.613699186s" podCreationTimestamp="2025-12-04 17:45:18 +0000 UTC" firstStartedPulling="2025-12-04 17:45:19.538214084 +0000 UTC m=+381.493575130" lastFinishedPulling="2025-12-04 17:45:20.984844785 +0000 UTC m=+382.940205841" observedRunningTime="2025-12-04 17:45:21.605331094 +0000 UTC m=+383.560692180" watchObservedRunningTime="2025-12-04 17:45:21.613699186 +0000 UTC m=+383.569060242" Dec 04 17:45:23 crc kubenswrapper[4733]: I1204 17:45:23.574156 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gb2qr" event={"ID":"d1d40421-3817-47d8-85c1-aeb819796b70","Type":"ContainerStarted","Data":"ff7c6214085835d802de429cbafbda2c0e293c5a1d0495ccffe104f061dd068d"} Dec 04 17:45:23 crc kubenswrapper[4733]: I1204 17:45:23.595904 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gb2qr" podStartSLOduration=3.129529312 podStartE2EDuration="4.59588388s" podCreationTimestamp="2025-12-04 17:45:19 +0000 UTC" firstStartedPulling="2025-12-04 17:45:20.552562706 +0000 UTC m=+382.507923772" lastFinishedPulling="2025-12-04 17:45:22.018917294 +0000 UTC m=+383.974278340" observedRunningTime="2025-12-04 17:45:23.593211046 +0000 UTC m=+385.548572092" watchObservedRunningTime="2025-12-04 17:45:23.59588388 +0000 UTC m=+385.551244966" Dec 04 17:45:26 crc kubenswrapper[4733]: I1204 17:45:26.260053 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:26 crc kubenswrapper[4733]: I1204 17:45:26.260412 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:26 crc kubenswrapper[4733]: I1204 17:45:26.297468 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:26 crc kubenswrapper[4733]: I1204 17:45:26.652027 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 17:45:27 crc kubenswrapper[4733]: I1204 17:45:27.273305 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:27 crc kubenswrapper[4733]: I1204 17:45:27.273873 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:27 crc kubenswrapper[4733]: I1204 17:45:27.339515 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:27 crc kubenswrapper[4733]: I1204 17:45:27.656348 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-khxrh" Dec 04 17:45:28 crc kubenswrapper[4733]: I1204 17:45:28.394607 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-llz25" Dec 04 17:45:28 crc kubenswrapper[4733]: I1204 17:45:28.447475 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rhg5t"] Dec 04 17:45:28 crc kubenswrapper[4733]: I1204 17:45:28.702066 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:28 crc kubenswrapper[4733]: I1204 17:45:28.702129 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:28 crc kubenswrapper[4733]: I1204 17:45:28.751845 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:29 crc kubenswrapper[4733]: I1204 17:45:29.660536 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cj6ms" Dec 04 17:45:29 crc kubenswrapper[4733]: I1204 17:45:29.694291 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:29 crc kubenswrapper[4733]: I1204 17:45:29.694332 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:29 crc kubenswrapper[4733]: I1204 17:45:29.764842 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:30 crc kubenswrapper[4733]: I1204 17:45:30.655212 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gb2qr" Dec 04 17:45:45 crc kubenswrapper[4733]: I1204 17:45:45.361846 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:45:45 crc kubenswrapper[4733]: I1204 17:45:45.362637 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:45:53 crc kubenswrapper[4733]: I1204 17:45:53.489703 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" podUID="91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" containerName="registry" containerID="cri-o://eb49efe3241be5daa80393c404745707c3d93bfd4668941f624d4d394c8e4db5" gracePeriod=30 Dec 04 17:45:53 crc kubenswrapper[4733]: E1204 17:45:53.613113 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91fd2fc9_3660_4fdc_b1a9_9f26572b44a4.slice/crio-eb49efe3241be5daa80393c404745707c3d93bfd4668941f624d4d394c8e4db5.scope\": RecentStats: unable to find data in memory cache]" Dec 04 17:45:54 crc kubenswrapper[4733]: I1204 17:45:54.769428 4733 generic.go:334] "Generic (PLEG): container finished" podID="91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" containerID="eb49efe3241be5daa80393c404745707c3d93bfd4668941f624d4d394c8e4db5" exitCode=0 Dec 04 17:45:54 crc kubenswrapper[4733]: I1204 17:45:54.769489 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" event={"ID":"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4","Type":"ContainerDied","Data":"eb49efe3241be5daa80393c404745707c3d93bfd4668941f624d4d394c8e4db5"} Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.764542 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.775092 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" event={"ID":"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4","Type":"ContainerDied","Data":"6a566a53da911dc6959bd5233f566fe0c52dd27c85757563ba51c9cedfc5ce3c"} Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.775137 4733 scope.go:117] "RemoveContainer" containerID="eb49efe3241be5daa80393c404745707c3d93bfd4668941f624d4d394c8e4db5" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.775216 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rhg5t" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.934216 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.934549 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-tls\") pod \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.934620 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-certificates\") pod \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.934690 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77rsq\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-kube-api-access-77rsq\") pod \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.934724 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-trusted-ca\") pod \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.934761 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-installation-pull-secrets\") pod \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.934834 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-bound-sa-token\") pod \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.934909 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-ca-trust-extracted\") pod \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\" (UID: \"91fd2fc9-3660-4fdc-b1a9-9f26572b44a4\") " Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.936083 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.936199 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.940634 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-kube-api-access-77rsq" (OuterVolumeSpecName: "kube-api-access-77rsq") pod "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4"). InnerVolumeSpecName "kube-api-access-77rsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.942101 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.945411 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.947116 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.949262 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:45:55 crc kubenswrapper[4733]: I1204 17:45:55.966664 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" (UID: "91fd2fc9-3660-4fdc-b1a9-9f26572b44a4"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.037041 4733 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.037088 4733 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.037108 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77rsq\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-kube-api-access-77rsq\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.037125 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.037143 4733 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.037159 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.037175 4733 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.125262 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rhg5t"] Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.135225 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rhg5t"] Dec 04 17:45:56 crc kubenswrapper[4733]: I1204 17:45:56.346582 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" path="/var/lib/kubelet/pods/91fd2fc9-3660-4fdc-b1a9-9f26572b44a4/volumes" Dec 04 17:46:15 crc kubenswrapper[4733]: I1204 17:46:15.362077 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:46:15 crc kubenswrapper[4733]: I1204 17:46:15.362836 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:46:15 crc kubenswrapper[4733]: I1204 17:46:15.362925 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:46:15 crc kubenswrapper[4733]: I1204 17:46:15.910639 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"adacb000dc8287db30045ce1678c37d9bee7575c1c73fa342ac484463a75e935"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 17:46:15 crc kubenswrapper[4733]: I1204 17:46:15.910757 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://adacb000dc8287db30045ce1678c37d9bee7575c1c73fa342ac484463a75e935" gracePeriod=600 Dec 04 17:46:16 crc kubenswrapper[4733]: I1204 17:46:16.919908 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="adacb000dc8287db30045ce1678c37d9bee7575c1c73fa342ac484463a75e935" exitCode=0 Dec 04 17:46:16 crc kubenswrapper[4733]: I1204 17:46:16.920066 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"adacb000dc8287db30045ce1678c37d9bee7575c1c73fa342ac484463a75e935"} Dec 04 17:46:16 crc kubenswrapper[4733]: I1204 17:46:16.920579 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"d83e5743bfe34b630102892e9bcb5c583d801852811e612149b74d0af61af29c"} Dec 04 17:46:16 crc kubenswrapper[4733]: I1204 17:46:16.920638 4733 scope.go:117] "RemoveContainer" containerID="60068448b03acff1f8962531c84f944f44e763d6e9226db9d5b80f2c9a520fb9" Dec 04 17:48:45 crc kubenswrapper[4733]: I1204 17:48:45.362183 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:48:45 crc kubenswrapper[4733]: I1204 17:48:45.362777 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:48:58 crc kubenswrapper[4733]: I1204 17:48:58.606553 4733 scope.go:117] "RemoveContainer" containerID="c733fe32ebf56d635aea87ebbf29acc3ed1f572e04ad703fa96e507f1da89e8f" Dec 04 17:49:15 crc kubenswrapper[4733]: I1204 17:49:15.361922 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:49:15 crc kubenswrapper[4733]: I1204 17:49:15.363861 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:49:45 crc kubenswrapper[4733]: I1204 17:49:45.362276 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:49:45 crc kubenswrapper[4733]: I1204 17:49:45.363262 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:49:45 crc kubenswrapper[4733]: I1204 17:49:45.364010 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:49:45 crc kubenswrapper[4733]: I1204 17:49:45.364900 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d83e5743bfe34b630102892e9bcb5c583d801852811e612149b74d0af61af29c"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 17:49:45 crc kubenswrapper[4733]: I1204 17:49:45.365000 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://d83e5743bfe34b630102892e9bcb5c583d801852811e612149b74d0af61af29c" gracePeriod=600 Dec 04 17:49:46 crc kubenswrapper[4733]: I1204 17:49:46.327343 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="d83e5743bfe34b630102892e9bcb5c583d801852811e612149b74d0af61af29c" exitCode=0 Dec 04 17:49:46 crc kubenswrapper[4733]: I1204 17:49:46.327531 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"d83e5743bfe34b630102892e9bcb5c583d801852811e612149b74d0af61af29c"} Dec 04 17:49:46 crc kubenswrapper[4733]: I1204 17:49:46.328074 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"ee5a68aff544c2577446d346027585de3353ade4bb096661850d64dabfc79180"} Dec 04 17:49:46 crc kubenswrapper[4733]: I1204 17:49:46.328109 4733 scope.go:117] "RemoveContainer" containerID="adacb000dc8287db30045ce1678c37d9bee7575c1c73fa342ac484463a75e935" Dec 04 17:51:37 crc kubenswrapper[4733]: I1204 17:51:37.069374 4733 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 17:51:45 crc kubenswrapper[4733]: I1204 17:51:45.361468 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:51:45 crc kubenswrapper[4733]: I1204 17:51:45.362524 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.089480 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m6jzk"] Dec 04 17:51:55 crc kubenswrapper[4733]: E1204 17:51:55.092185 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" containerName="registry" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.092398 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" containerName="registry" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.092719 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="91fd2fc9-3660-4fdc-b1a9-9f26572b44a4" containerName="registry" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.094855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.102141 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-catalog-content\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.102229 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff5fm\" (UniqueName: \"kubernetes.io/projected/38e50b6c-5091-406c-acec-c21d410853a0-kube-api-access-ff5fm\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.102443 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-utilities\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.129277 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m6jzk"] Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.204052 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-utilities\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.204345 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-catalog-content\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.204481 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff5fm\" (UniqueName: \"kubernetes.io/projected/38e50b6c-5091-406c-acec-c21d410853a0-kube-api-access-ff5fm\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.204676 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-utilities\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.204752 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-catalog-content\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.230196 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff5fm\" (UniqueName: \"kubernetes.io/projected/38e50b6c-5091-406c-acec-c21d410853a0-kube-api-access-ff5fm\") pod \"community-operators-m6jzk\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.417070 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:51:55 crc kubenswrapper[4733]: I1204 17:51:55.680053 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m6jzk"] Dec 04 17:51:56 crc kubenswrapper[4733]: I1204 17:51:56.190986 4733 generic.go:334] "Generic (PLEG): container finished" podID="38e50b6c-5091-406c-acec-c21d410853a0" containerID="beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24" exitCode=0 Dec 04 17:51:56 crc kubenswrapper[4733]: I1204 17:51:56.191102 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6jzk" event={"ID":"38e50b6c-5091-406c-acec-c21d410853a0","Type":"ContainerDied","Data":"beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24"} Dec 04 17:51:56 crc kubenswrapper[4733]: I1204 17:51:56.191392 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6jzk" event={"ID":"38e50b6c-5091-406c-acec-c21d410853a0","Type":"ContainerStarted","Data":"86544dd105fe66d3a46436cfccf5b942a32494db80255b65297f4dcba64d8d93"} Dec 04 17:51:56 crc kubenswrapper[4733]: I1204 17:51:56.194631 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 17:51:57 crc kubenswrapper[4733]: I1204 17:51:57.197949 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6jzk" event={"ID":"38e50b6c-5091-406c-acec-c21d410853a0","Type":"ContainerStarted","Data":"fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998"} Dec 04 17:51:58 crc kubenswrapper[4733]: I1204 17:51:58.208313 4733 generic.go:334] "Generic (PLEG): container finished" podID="38e50b6c-5091-406c-acec-c21d410853a0" containerID="fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998" exitCode=0 Dec 04 17:51:58 crc kubenswrapper[4733]: I1204 17:51:58.208396 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6jzk" event={"ID":"38e50b6c-5091-406c-acec-c21d410853a0","Type":"ContainerDied","Data":"fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998"} Dec 04 17:51:59 crc kubenswrapper[4733]: I1204 17:51:59.227778 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6jzk" event={"ID":"38e50b6c-5091-406c-acec-c21d410853a0","Type":"ContainerStarted","Data":"a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9"} Dec 04 17:51:59 crc kubenswrapper[4733]: I1204 17:51:59.262187 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m6jzk" podStartSLOduration=1.674327181 podStartE2EDuration="4.262161172s" podCreationTimestamp="2025-12-04 17:51:55 +0000 UTC" firstStartedPulling="2025-12-04 17:51:56.194268936 +0000 UTC m=+778.149630012" lastFinishedPulling="2025-12-04 17:51:58.782102917 +0000 UTC m=+780.737464003" observedRunningTime="2025-12-04 17:51:59.255118809 +0000 UTC m=+781.210479925" watchObservedRunningTime="2025-12-04 17:51:59.262161172 +0000 UTC m=+781.217522248" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.248251 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qp9sx"] Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.249875 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovn-controller" containerID="cri-o://440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc" gracePeriod=30 Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.250138 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="sbdb" containerID="cri-o://d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320" gracePeriod=30 Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.250174 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9" gracePeriod=30 Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.250246 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kube-rbac-proxy-node" containerID="cri-o://ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291" gracePeriod=30 Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.250269 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="nbdb" containerID="cri-o://b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b" gracePeriod=30 Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.250472 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="northd" containerID="cri-o://fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2" gracePeriod=30 Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.250363 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovn-acl-logging" containerID="cri-o://18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d" gracePeriod=30 Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.350364 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" containerID="cri-o://a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" gracePeriod=30 Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.417506 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.418412 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.492076 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.592292 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/3.log" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.594427 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovn-acl-logging/0.log" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.594858 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovn-controller/0.log" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.595417 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.641767 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8p74n"] Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642266 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovn-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642282 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovn-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642296 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642305 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642316 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="nbdb" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642325 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="nbdb" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642335 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kube-rbac-proxy-node" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642343 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kube-rbac-proxy-node" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642355 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642363 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642374 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="sbdb" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642381 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="sbdb" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642389 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642397 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642405 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642413 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642425 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kubecfg-setup" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642433 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kubecfg-setup" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642446 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642454 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642466 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovn-acl-logging" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642474 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovn-acl-logging" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642488 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="northd" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642495 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="northd" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642600 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642612 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642622 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642631 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642638 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovn-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642647 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642659 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="kube-rbac-proxy-node" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642670 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="sbdb" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642678 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="northd" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642689 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="nbdb" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642699 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovn-acl-logging" Dec 04 17:52:05 crc kubenswrapper[4733]: E1204 17:52:05.642821 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642831 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.642935 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerName="ovnkube-controller" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.644762 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765080 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-systemd\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765143 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhdsl\" (UniqueName: \"kubernetes.io/projected/067d57d5-c6ec-463d-b1a0-ec5e8c686453-kube-api-access-vhdsl\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765169 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-var-lib-openvswitch\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765202 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-ovn\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765245 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-env-overrides\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765256 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765287 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovn-node-metrics-cert\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765340 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765357 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-netd\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765399 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-slash\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765431 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-systemd-units\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765460 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-netns\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765516 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-etc-openvswitch\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765543 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-config\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765562 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-kubelet\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765587 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-var-lib-cni-networks-ovn-kubernetes\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765613 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-bin\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765634 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-log-socket\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765651 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-node-log\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765685 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765689 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-ovn-kubernetes\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765743 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765757 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-openvswitch\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765811 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-script-lib\") pod \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\" (UID: \"067d57d5-c6ec-463d-b1a0-ec5e8c686453\") " Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765897 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-ovn\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765925 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-run-ovn-kubernetes\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765963 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-env-overrides\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.765989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-ovnkube-script-lib\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766008 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-systemd-units\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766029 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj624\" (UniqueName: \"kubernetes.io/projected/db2b0de0-462c-4726-af72-a54f3fd0349a-kube-api-access-bj624\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766050 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/db2b0de0-462c-4726-af72-a54f3fd0349a-ovn-node-metrics-cert\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766075 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-kubelet\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766110 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-cni-netd\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766127 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-node-log\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766156 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-ovnkube-config\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766180 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-var-lib-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766209 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-cni-bin\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766233 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766258 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-slash\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766282 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-run-netns\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766208 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766231 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766319 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766250 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766250 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-log-socket" (OuterVolumeSpecName: "log-socket") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766277 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766287 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766333 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766321 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-node-log" (OuterVolumeSpecName: "node-log") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766380 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766294 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-slash" (OuterVolumeSpecName: "host-slash") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766420 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-etc-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766584 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766723 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-log-socket\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766760 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-systemd\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766759 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766866 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766913 4733 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-slash\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.766940 4733 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.767452 4733 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.767761 4733 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.767780 4733 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.767835 4733 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.767857 4733 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.767876 4733 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-log-socket\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780360 4733 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-node-log\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780404 4733 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780435 4733 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780451 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780464 4733 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780480 4733 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780498 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780511 4733 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.780897 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/067d57d5-c6ec-463d-b1a0-ec5e8c686453-kube-api-access-vhdsl" (OuterVolumeSpecName: "kube-api-access-vhdsl") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "kube-api-access-vhdsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.781136 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.784467 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "067d57d5-c6ec-463d-b1a0-ec5e8c686453" (UID: "067d57d5-c6ec-463d-b1a0-ec5e8c686453"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882203 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-ovnkube-script-lib\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882248 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-systemd-units\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882269 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj624\" (UniqueName: \"kubernetes.io/projected/db2b0de0-462c-4726-af72-a54f3fd0349a-kube-api-access-bj624\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882288 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/db2b0de0-462c-4726-af72-a54f3fd0349a-ovn-node-metrics-cert\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-kubelet\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882345 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-cni-netd\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882363 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-node-log\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-ovnkube-config\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882412 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-var-lib-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882436 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-cni-bin\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882453 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882475 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-slash\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882500 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-run-netns\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882529 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-etc-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882556 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882585 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-log-socket\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882607 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-systemd\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882633 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-ovn\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882656 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-run-ovn-kubernetes\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882688 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-env-overrides\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882733 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882749 4733 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/067d57d5-c6ec-463d-b1a0-ec5e8c686453-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882763 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhdsl\" (UniqueName: \"kubernetes.io/projected/067d57d5-c6ec-463d-b1a0-ec5e8c686453-kube-api-access-vhdsl\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.882777 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/067d57d5-c6ec-463d-b1a0-ec5e8c686453-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.883392 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-env-overrides\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.883878 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-ovnkube-script-lib\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.883931 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-systemd-units\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.883964 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-log-socket\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.883990 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-systemd\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884017 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-ovn\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884007 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884066 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-slash\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.883881 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-cni-bin\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884044 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-run-ovn-kubernetes\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884117 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-run-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884119 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-run-netns\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884149 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-etc-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884171 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-cni-netd\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884243 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-var-lib-openvswitch\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884285 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-node-log\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884314 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/db2b0de0-462c-4726-af72-a54f3fd0349a-host-kubelet\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.884681 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/db2b0de0-462c-4726-af72-a54f3fd0349a-ovnkube-config\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.889147 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/db2b0de0-462c-4726-af72-a54f3fd0349a-ovn-node-metrics-cert\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.914481 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj624\" (UniqueName: \"kubernetes.io/projected/db2b0de0-462c-4726-af72-a54f3fd0349a-kube-api-access-bj624\") pod \"ovnkube-node-8p74n\" (UID: \"db2b0de0-462c-4726-af72-a54f3fd0349a\") " pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:05 crc kubenswrapper[4733]: I1204 17:52:05.958958 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.277913 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovnkube-controller/3.log" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.281502 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovn-acl-logging/0.log" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.282194 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qp9sx_067d57d5-c6ec-463d-b1a0-ec5e8c686453/ovn-controller/0.log" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.282830 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" exitCode=0 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.282895 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320" exitCode=0 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.282924 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b" exitCode=0 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.282949 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2" exitCode=0 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.282974 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9" exitCode=0 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.282996 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291" exitCode=0 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283013 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d" exitCode=143 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283019 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283037 4733 generic.go:334] "Generic (PLEG): container finished" podID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" containerID="440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc" exitCode=143 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.282934 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283121 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283150 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283172 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283192 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283212 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283233 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283251 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283263 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283275 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283286 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283297 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283310 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283321 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283333 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283346 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283363 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283376 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283387 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283397 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283408 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283296 4733 scope.go:117] "RemoveContainer" containerID="a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283418 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283545 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283558 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283570 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283582 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283599 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283617 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283629 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283641 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283653 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283664 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283675 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283686 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283697 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283708 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283719 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283734 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qp9sx" event={"ID":"067d57d5-c6ec-463d-b1a0-ec5e8c686453","Type":"ContainerDied","Data":"f4aa35112d541e6a6bc44ab54969cdb9311f1517f34e17ae8ca00f7684347f60"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283749 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283763 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283773 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283784 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283823 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283835 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283846 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283856 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283867 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.283878 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.285044 4733 generic.go:334] "Generic (PLEG): container finished" podID="db2b0de0-462c-4726-af72-a54f3fd0349a" containerID="9acfc9ac6b4d9ebf58cc4b7adfca74693108b54dc0adcc9f081b78242eafffba" exitCode=0 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.285121 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerDied","Data":"9acfc9ac6b4d9ebf58cc4b7adfca74693108b54dc0adcc9f081b78242eafffba"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.285160 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"75701c8593df7d56b9da1a84424325e740018e9f97e3ab158a6289096d6ad662"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.288723 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/2.log" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.289484 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/1.log" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.289523 4733 generic.go:334] "Generic (PLEG): container finished" podID="829fda7a-6a96-4007-b2f5-0be03782a764" containerID="05f1cdc7acc55339ce7a4767f128c0cfbb18eb7c3649d4d49c2ac1680976c7bc" exitCode=2 Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.289627 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-472m8" event={"ID":"829fda7a-6a96-4007-b2f5-0be03782a764","Type":"ContainerDied","Data":"05f1cdc7acc55339ce7a4767f128c0cfbb18eb7c3649d4d49c2ac1680976c7bc"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.289701 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886"} Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.290251 4733 scope.go:117] "RemoveContainer" containerID="05f1cdc7acc55339ce7a4767f128c0cfbb18eb7c3649d4d49c2ac1680976c7bc" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.343889 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.379196 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.379264 4733 scope.go:117] "RemoveContainer" containerID="d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.385960 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qp9sx"] Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.389405 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qp9sx"] Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.436603 4733 scope.go:117] "RemoveContainer" containerID="b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.438900 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m6jzk"] Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.451235 4733 scope.go:117] "RemoveContainer" containerID="fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.466212 4733 scope.go:117] "RemoveContainer" containerID="cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.492951 4733 scope.go:117] "RemoveContainer" containerID="ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.508502 4733 scope.go:117] "RemoveContainer" containerID="18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.529151 4733 scope.go:117] "RemoveContainer" containerID="440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.561518 4733 scope.go:117] "RemoveContainer" containerID="95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.581595 4733 scope.go:117] "RemoveContainer" containerID="a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.583722 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": container with ID starting with a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a not found: ID does not exist" containerID="a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.583755 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} err="failed to get container status \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": rpc error: code = NotFound desc = could not find container \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": container with ID starting with a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.583786 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.584076 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": container with ID starting with 7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650 not found: ID does not exist" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.584104 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} err="failed to get container status \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": rpc error: code = NotFound desc = could not find container \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": container with ID starting with 7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.584123 4733 scope.go:117] "RemoveContainer" containerID="d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.584362 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": container with ID starting with d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320 not found: ID does not exist" containerID="d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.584392 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} err="failed to get container status \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": rpc error: code = NotFound desc = could not find container \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": container with ID starting with d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.584413 4733 scope.go:117] "RemoveContainer" containerID="b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.584727 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": container with ID starting with b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b not found: ID does not exist" containerID="b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.584752 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} err="failed to get container status \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": rpc error: code = NotFound desc = could not find container \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": container with ID starting with b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.584770 4733 scope.go:117] "RemoveContainer" containerID="fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.585281 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": container with ID starting with fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2 not found: ID does not exist" containerID="fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.585311 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} err="failed to get container status \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": rpc error: code = NotFound desc = could not find container \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": container with ID starting with fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.585328 4733 scope.go:117] "RemoveContainer" containerID="cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.585627 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": container with ID starting with cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9 not found: ID does not exist" containerID="cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.585650 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} err="failed to get container status \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": rpc error: code = NotFound desc = could not find container \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": container with ID starting with cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.585666 4733 scope.go:117] "RemoveContainer" containerID="ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.586007 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": container with ID starting with ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291 not found: ID does not exist" containerID="ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.586033 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} err="failed to get container status \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": rpc error: code = NotFound desc = could not find container \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": container with ID starting with ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.586051 4733 scope.go:117] "RemoveContainer" containerID="18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.586331 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": container with ID starting with 18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d not found: ID does not exist" containerID="18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.586359 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} err="failed to get container status \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": rpc error: code = NotFound desc = could not find container \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": container with ID starting with 18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.586375 4733 scope.go:117] "RemoveContainer" containerID="440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.586668 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": container with ID starting with 440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc not found: ID does not exist" containerID="440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.586691 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} err="failed to get container status \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": rpc error: code = NotFound desc = could not find container \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": container with ID starting with 440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.586711 4733 scope.go:117] "RemoveContainer" containerID="95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18" Dec 04 17:52:06 crc kubenswrapper[4733]: E1204 17:52:06.587099 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": container with ID starting with 95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18 not found: ID does not exist" containerID="95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.587123 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} err="failed to get container status \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": rpc error: code = NotFound desc = could not find container \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": container with ID starting with 95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.587139 4733 scope.go:117] "RemoveContainer" containerID="a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.587419 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} err="failed to get container status \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": rpc error: code = NotFound desc = could not find container \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": container with ID starting with a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.587441 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.587681 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} err="failed to get container status \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": rpc error: code = NotFound desc = could not find container \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": container with ID starting with 7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.587702 4733 scope.go:117] "RemoveContainer" containerID="d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.587977 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} err="failed to get container status \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": rpc error: code = NotFound desc = could not find container \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": container with ID starting with d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.587998 4733 scope.go:117] "RemoveContainer" containerID="b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.588240 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} err="failed to get container status \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": rpc error: code = NotFound desc = could not find container \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": container with ID starting with b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.588261 4733 scope.go:117] "RemoveContainer" containerID="fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.588527 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} err="failed to get container status \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": rpc error: code = NotFound desc = could not find container \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": container with ID starting with fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.588548 4733 scope.go:117] "RemoveContainer" containerID="cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.588824 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} err="failed to get container status \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": rpc error: code = NotFound desc = could not find container \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": container with ID starting with cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.588849 4733 scope.go:117] "RemoveContainer" containerID="ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.589109 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} err="failed to get container status \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": rpc error: code = NotFound desc = could not find container \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": container with ID starting with ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.589131 4733 scope.go:117] "RemoveContainer" containerID="18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.589391 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} err="failed to get container status \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": rpc error: code = NotFound desc = could not find container \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": container with ID starting with 18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.589412 4733 scope.go:117] "RemoveContainer" containerID="440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.589671 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} err="failed to get container status \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": rpc error: code = NotFound desc = could not find container \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": container with ID starting with 440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.589692 4733 scope.go:117] "RemoveContainer" containerID="95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.590194 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} err="failed to get container status \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": rpc error: code = NotFound desc = could not find container \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": container with ID starting with 95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.590215 4733 scope.go:117] "RemoveContainer" containerID="a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.590758 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} err="failed to get container status \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": rpc error: code = NotFound desc = could not find container \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": container with ID starting with a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.590779 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.591081 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} err="failed to get container status \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": rpc error: code = NotFound desc = could not find container \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": container with ID starting with 7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.591103 4733 scope.go:117] "RemoveContainer" containerID="d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.591501 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} err="failed to get container status \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": rpc error: code = NotFound desc = could not find container \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": container with ID starting with d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.591544 4733 scope.go:117] "RemoveContainer" containerID="b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.592268 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} err="failed to get container status \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": rpc error: code = NotFound desc = could not find container \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": container with ID starting with b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.592315 4733 scope.go:117] "RemoveContainer" containerID="fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.595490 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} err="failed to get container status \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": rpc error: code = NotFound desc = could not find container \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": container with ID starting with fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.595541 4733 scope.go:117] "RemoveContainer" containerID="cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.596075 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} err="failed to get container status \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": rpc error: code = NotFound desc = could not find container \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": container with ID starting with cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.596109 4733 scope.go:117] "RemoveContainer" containerID="ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.596534 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} err="failed to get container status \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": rpc error: code = NotFound desc = could not find container \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": container with ID starting with ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.596575 4733 scope.go:117] "RemoveContainer" containerID="18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.596938 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} err="failed to get container status \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": rpc error: code = NotFound desc = could not find container \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": container with ID starting with 18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.596969 4733 scope.go:117] "RemoveContainer" containerID="440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.597246 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} err="failed to get container status \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": rpc error: code = NotFound desc = could not find container \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": container with ID starting with 440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.597269 4733 scope.go:117] "RemoveContainer" containerID="95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.597614 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} err="failed to get container status \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": rpc error: code = NotFound desc = could not find container \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": container with ID starting with 95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.597662 4733 scope.go:117] "RemoveContainer" containerID="a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.598017 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} err="failed to get container status \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": rpc error: code = NotFound desc = could not find container \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": container with ID starting with a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.598047 4733 scope.go:117] "RemoveContainer" containerID="7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.598299 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650"} err="failed to get container status \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": rpc error: code = NotFound desc = could not find container \"7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650\": container with ID starting with 7e366158843c29b672d362d03fefd815333bec00162398c6526cbc373a4c2650 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.598324 4733 scope.go:117] "RemoveContainer" containerID="d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.598599 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320"} err="failed to get container status \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": rpc error: code = NotFound desc = could not find container \"d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320\": container with ID starting with d58a33c788ff305f584bf213b9e45ed9b75ad84d80d4aef2acdbd83bb40ac320 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.598620 4733 scope.go:117] "RemoveContainer" containerID="b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.598925 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b"} err="failed to get container status \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": rpc error: code = NotFound desc = could not find container \"b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b\": container with ID starting with b3832e920e28a2db1b49c7dfa9fa8183bca15495c38214e22eddc6b1d3c11f8b not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.598955 4733 scope.go:117] "RemoveContainer" containerID="fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.599225 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2"} err="failed to get container status \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": rpc error: code = NotFound desc = could not find container \"fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2\": container with ID starting with fa62912d00cbc297ee8338af9b0399c851cc024adc8280b311be5b8656f8e9f2 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.599243 4733 scope.go:117] "RemoveContainer" containerID="cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.599523 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9"} err="failed to get container status \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": rpc error: code = NotFound desc = could not find container \"cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9\": container with ID starting with cf9da57ec76f18dcab86159bd91d9e9ee9a96540783b2a889ca94d1cd649fda9 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.599553 4733 scope.go:117] "RemoveContainer" containerID="ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.599857 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291"} err="failed to get container status \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": rpc error: code = NotFound desc = could not find container \"ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291\": container with ID starting with ff30008d4dc65feceea9af787c77b854dcb31712fd7d8006472ce156ab848291 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.599877 4733 scope.go:117] "RemoveContainer" containerID="18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.600289 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d"} err="failed to get container status \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": rpc error: code = NotFound desc = could not find container \"18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d\": container with ID starting with 18db591880e5384a929651251f29f82e8faa2ada6aed03a41daa840a2f3e239d not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.600306 4733 scope.go:117] "RemoveContainer" containerID="440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.600562 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc"} err="failed to get container status \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": rpc error: code = NotFound desc = could not find container \"440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc\": container with ID starting with 440d46f1a0ed10392ea4aba9cc7626f0b9bc15515b45885ebee86d8111d5c7dc not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.600585 4733 scope.go:117] "RemoveContainer" containerID="95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.600924 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18"} err="failed to get container status \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": rpc error: code = NotFound desc = could not find container \"95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18\": container with ID starting with 95e65d1c40aedf1a26b6f6240e70adcbd01be3562b1e623c3fa42533448d3c18 not found: ID does not exist" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.600944 4733 scope.go:117] "RemoveContainer" containerID="a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a" Dec 04 17:52:06 crc kubenswrapper[4733]: I1204 17:52:06.601165 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a"} err="failed to get container status \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": rpc error: code = NotFound desc = could not find container \"a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a\": container with ID starting with a8fb799f6c1d25558d9464e1f65cd9594c8dd9df9985984268a2400e4e2f856a not found: ID does not exist" Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.303645 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"a078a77dc6aa90840a5b223aadbda4318a9ccea8068f64047fe549456cb62f00"} Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.304853 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"8809df99e739b081cf5d5d7e7b97d9c3dd937533cdc6012703c2c8436e29fe82"} Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.304967 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"d8e5b2497a76494e1d928f72da2796fc03d279e0ca7bff49f00e9bba1af5a503"} Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.305036 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"d870f2d7d08c36aac73e4d4b28e6c26909f839b047f70750829a92de6b1f8138"} Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.305127 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"336cd4749c2f898081d13966edab7c27451a01a9358c98f06f08523b2b337481"} Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.305190 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"ffc12111e859e2e51a108159f4150450a0f226f1e0823d20e6b8174291b53f7a"} Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.306417 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/2.log" Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.307122 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/1.log" Dec 04 17:52:07 crc kubenswrapper[4733]: I1204 17:52:07.307624 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-472m8" event={"ID":"829fda7a-6a96-4007-b2f5-0be03782a764","Type":"ContainerStarted","Data":"0c64a41e8ab190df0d205b043ab1eba0bef91bdc872b3e3d7a70cacffbbb5d43"} Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.316397 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m6jzk" podUID="38e50b6c-5091-406c-acec-c21d410853a0" containerName="registry-server" containerID="cri-o://a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9" gracePeriod=2 Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.345858 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="067d57d5-c6ec-463d-b1a0-ec5e8c686453" path="/var/lib/kubelet/pods/067d57d5-c6ec-463d-b1a0-ec5e8c686453/volumes" Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.510782 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.630539 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-catalog-content\") pod \"38e50b6c-5091-406c-acec-c21d410853a0\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.630624 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff5fm\" (UniqueName: \"kubernetes.io/projected/38e50b6c-5091-406c-acec-c21d410853a0-kube-api-access-ff5fm\") pod \"38e50b6c-5091-406c-acec-c21d410853a0\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.630727 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-utilities\") pod \"38e50b6c-5091-406c-acec-c21d410853a0\" (UID: \"38e50b6c-5091-406c-acec-c21d410853a0\") " Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.632665 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-utilities" (OuterVolumeSpecName: "utilities") pod "38e50b6c-5091-406c-acec-c21d410853a0" (UID: "38e50b6c-5091-406c-acec-c21d410853a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.642578 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38e50b6c-5091-406c-acec-c21d410853a0-kube-api-access-ff5fm" (OuterVolumeSpecName: "kube-api-access-ff5fm") pod "38e50b6c-5091-406c-acec-c21d410853a0" (UID: "38e50b6c-5091-406c-acec-c21d410853a0"). InnerVolumeSpecName "kube-api-access-ff5fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.681752 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38e50b6c-5091-406c-acec-c21d410853a0" (UID: "38e50b6c-5091-406c-acec-c21d410853a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.732417 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.732456 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff5fm\" (UniqueName: \"kubernetes.io/projected/38e50b6c-5091-406c-acec-c21d410853a0-kube-api-access-ff5fm\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:08 crc kubenswrapper[4733]: I1204 17:52:08.732471 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e50b6c-5091-406c-acec-c21d410853a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.324214 4733 generic.go:334] "Generic (PLEG): container finished" podID="38e50b6c-5091-406c-acec-c21d410853a0" containerID="a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9" exitCode=0 Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.324290 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6jzk" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.324304 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6jzk" event={"ID":"38e50b6c-5091-406c-acec-c21d410853a0","Type":"ContainerDied","Data":"a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9"} Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.324733 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6jzk" event={"ID":"38e50b6c-5091-406c-acec-c21d410853a0","Type":"ContainerDied","Data":"86544dd105fe66d3a46436cfccf5b942a32494db80255b65297f4dcba64d8d93"} Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.324772 4733 scope.go:117] "RemoveContainer" containerID="a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.344302 4733 scope.go:117] "RemoveContainer" containerID="fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.366448 4733 scope.go:117] "RemoveContainer" containerID="beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.375384 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m6jzk"] Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.378654 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m6jzk"] Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.398008 4733 scope.go:117] "RemoveContainer" containerID="a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9" Dec 04 17:52:09 crc kubenswrapper[4733]: E1204 17:52:09.398337 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9\": container with ID starting with a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9 not found: ID does not exist" containerID="a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.398365 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9"} err="failed to get container status \"a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9\": rpc error: code = NotFound desc = could not find container \"a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9\": container with ID starting with a17308ce8dc206c57d43b7a2ec100e40c47bd4514ef3cf89618ff8bc8784b6b9 not found: ID does not exist" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.398384 4733 scope.go:117] "RemoveContainer" containerID="fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998" Dec 04 17:52:09 crc kubenswrapper[4733]: E1204 17:52:09.398727 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998\": container with ID starting with fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998 not found: ID does not exist" containerID="fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.398832 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998"} err="failed to get container status \"fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998\": rpc error: code = NotFound desc = could not find container \"fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998\": container with ID starting with fa695a517b0746dc9657441462ebb17a9419902abccf1383e6d10bee339d0998 not found: ID does not exist" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.398892 4733 scope.go:117] "RemoveContainer" containerID="beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24" Dec 04 17:52:09 crc kubenswrapper[4733]: E1204 17:52:09.399273 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24\": container with ID starting with beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24 not found: ID does not exist" containerID="beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24" Dec 04 17:52:09 crc kubenswrapper[4733]: I1204 17:52:09.399300 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24"} err="failed to get container status \"beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24\": rpc error: code = NotFound desc = could not find container \"beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24\": container with ID starting with beb6b624a8c0bd3ff65d21b87da73bd2dc5e21946dbf163799b92c56ba545b24 not found: ID does not exist" Dec 04 17:52:09 crc kubenswrapper[4733]: E1204 17:52:09.428327 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38e50b6c_5091_406c_acec_c21d410853a0.slice/crio-86544dd105fe66d3a46436cfccf5b942a32494db80255b65297f4dcba64d8d93\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38e50b6c_5091_406c_acec_c21d410853a0.slice\": RecentStats: unable to find data in memory cache]" Dec 04 17:52:10 crc kubenswrapper[4733]: I1204 17:52:10.350215 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38e50b6c-5091-406c-acec-c21d410853a0" path="/var/lib/kubelet/pods/38e50b6c-5091-406c-acec-c21d410853a0/volumes" Dec 04 17:52:10 crc kubenswrapper[4733]: I1204 17:52:10.351723 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"aa0832075f9b9333182077a4c67cc117b30de10363e193038e8353ed90c3ffd2"} Dec 04 17:52:12 crc kubenswrapper[4733]: I1204 17:52:12.355098 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" event={"ID":"db2b0de0-462c-4726-af72-a54f3fd0349a","Type":"ContainerStarted","Data":"706beac6a414abe1b1fdfa2fc37f55987ddcdd7655ac00a3e86b2caa907e3e88"} Dec 04 17:52:12 crc kubenswrapper[4733]: I1204 17:52:12.355995 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:12 crc kubenswrapper[4733]: I1204 17:52:12.356017 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:12 crc kubenswrapper[4733]: I1204 17:52:12.356029 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:12 crc kubenswrapper[4733]: I1204 17:52:12.390210 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" podStartSLOduration=7.390193245 podStartE2EDuration="7.390193245s" podCreationTimestamp="2025-12-04 17:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:52:12.388307963 +0000 UTC m=+794.343669029" watchObservedRunningTime="2025-12-04 17:52:12.390193245 +0000 UTC m=+794.345554301" Dec 04 17:52:12 crc kubenswrapper[4733]: I1204 17:52:12.396553 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:12 crc kubenswrapper[4733]: I1204 17:52:12.399451 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.155233 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-t5jhj"] Dec 04 17:52:14 crc kubenswrapper[4733]: E1204 17:52:14.155466 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38e50b6c-5091-406c-acec-c21d410853a0" containerName="registry-server" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.155483 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="38e50b6c-5091-406c-acec-c21d410853a0" containerName="registry-server" Dec 04 17:52:14 crc kubenswrapper[4733]: E1204 17:52:14.155498 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38e50b6c-5091-406c-acec-c21d410853a0" containerName="extract-content" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.155507 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="38e50b6c-5091-406c-acec-c21d410853a0" containerName="extract-content" Dec 04 17:52:14 crc kubenswrapper[4733]: E1204 17:52:14.155522 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38e50b6c-5091-406c-acec-c21d410853a0" containerName="extract-utilities" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.155529 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="38e50b6c-5091-406c-acec-c21d410853a0" containerName="extract-utilities" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.155636 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="38e50b6c-5091-406c-acec-c21d410853a0" containerName="registry-server" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.156087 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.158763 4733 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-r9cpb" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.159109 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.160332 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.160564 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.173121 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-t5jhj"] Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.202741 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5addc135-8f0f-4edd-ad29-76f3fd340827-node-mnt\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.202920 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz7nt\" (UniqueName: \"kubernetes.io/projected/5addc135-8f0f-4edd-ad29-76f3fd340827-kube-api-access-vz7nt\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.203017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5addc135-8f0f-4edd-ad29-76f3fd340827-crc-storage\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.304480 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz7nt\" (UniqueName: \"kubernetes.io/projected/5addc135-8f0f-4edd-ad29-76f3fd340827-kube-api-access-vz7nt\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.304922 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5addc135-8f0f-4edd-ad29-76f3fd340827-crc-storage\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.305018 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5addc135-8f0f-4edd-ad29-76f3fd340827-node-mnt\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.305333 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5addc135-8f0f-4edd-ad29-76f3fd340827-node-mnt\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.307963 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5addc135-8f0f-4edd-ad29-76f3fd340827-crc-storage\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.340297 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz7nt\" (UniqueName: \"kubernetes.io/projected/5addc135-8f0f-4edd-ad29-76f3fd340827-kube-api-access-vz7nt\") pod \"crc-storage-crc-t5jhj\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: I1204 17:52:14.485444 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: E1204 17:52:14.527630 4733 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-t5jhj_crc-storage_5addc135-8f0f-4edd-ad29-76f3fd340827_0(a5d574827656a68e515463304c19ed39d1e736dde76ae4476e231a228d13ea78): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 17:52:14 crc kubenswrapper[4733]: E1204 17:52:14.527775 4733 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-t5jhj_crc-storage_5addc135-8f0f-4edd-ad29-76f3fd340827_0(a5d574827656a68e515463304c19ed39d1e736dde76ae4476e231a228d13ea78): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: E1204 17:52:14.527872 4733 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-t5jhj_crc-storage_5addc135-8f0f-4edd-ad29-76f3fd340827_0(a5d574827656a68e515463304c19ed39d1e736dde76ae4476e231a228d13ea78): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:14 crc kubenswrapper[4733]: E1204 17:52:14.527978 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-t5jhj_crc-storage(5addc135-8f0f-4edd-ad29-76f3fd340827)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-t5jhj_crc-storage(5addc135-8f0f-4edd-ad29-76f3fd340827)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-t5jhj_crc-storage_5addc135-8f0f-4edd-ad29-76f3fd340827_0(a5d574827656a68e515463304c19ed39d1e736dde76ae4476e231a228d13ea78): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-t5jhj" podUID="5addc135-8f0f-4edd-ad29-76f3fd340827" Dec 04 17:52:15 crc kubenswrapper[4733]: I1204 17:52:15.362397 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:52:15 crc kubenswrapper[4733]: I1204 17:52:15.362468 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:52:15 crc kubenswrapper[4733]: I1204 17:52:15.374839 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:15 crc kubenswrapper[4733]: I1204 17:52:15.375545 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:15 crc kubenswrapper[4733]: E1204 17:52:15.421961 4733 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-t5jhj_crc-storage_5addc135-8f0f-4edd-ad29-76f3fd340827_0(17be1dcf8e896ee537af260770a2affde5c896742a6faf1ff5e98438c31de2f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 17:52:15 crc kubenswrapper[4733]: E1204 17:52:15.422030 4733 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-t5jhj_crc-storage_5addc135-8f0f-4edd-ad29-76f3fd340827_0(17be1dcf8e896ee537af260770a2affde5c896742a6faf1ff5e98438c31de2f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:15 crc kubenswrapper[4733]: E1204 17:52:15.422059 4733 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-t5jhj_crc-storage_5addc135-8f0f-4edd-ad29-76f3fd340827_0(17be1dcf8e896ee537af260770a2affde5c896742a6faf1ff5e98438c31de2f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:15 crc kubenswrapper[4733]: E1204 17:52:15.422116 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-t5jhj_crc-storage(5addc135-8f0f-4edd-ad29-76f3fd340827)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-t5jhj_crc-storage(5addc135-8f0f-4edd-ad29-76f3fd340827)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-t5jhj_crc-storage_5addc135-8f0f-4edd-ad29-76f3fd340827_0(17be1dcf8e896ee537af260770a2affde5c896742a6faf1ff5e98438c31de2f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-t5jhj" podUID="5addc135-8f0f-4edd-ad29-76f3fd340827" Dec 04 17:52:26 crc kubenswrapper[4733]: I1204 17:52:26.335155 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:26 crc kubenswrapper[4733]: I1204 17:52:26.337258 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:26 crc kubenswrapper[4733]: I1204 17:52:26.623222 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-t5jhj"] Dec 04 17:52:27 crc kubenswrapper[4733]: I1204 17:52:27.467637 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-t5jhj" event={"ID":"5addc135-8f0f-4edd-ad29-76f3fd340827","Type":"ContainerStarted","Data":"5ab60af3821462e94a340acfbff67263f71ec039fac233f01175612249f3c5db"} Dec 04 17:52:28 crc kubenswrapper[4733]: I1204 17:52:28.478977 4733 generic.go:334] "Generic (PLEG): container finished" podID="5addc135-8f0f-4edd-ad29-76f3fd340827" containerID="0ed86f8827380f4b3fae7dd124841039d9a30de9d56cf47428336819a2069c00" exitCode=0 Dec 04 17:52:28 crc kubenswrapper[4733]: I1204 17:52:28.479308 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-t5jhj" event={"ID":"5addc135-8f0f-4edd-ad29-76f3fd340827","Type":"ContainerDied","Data":"0ed86f8827380f4b3fae7dd124841039d9a30de9d56cf47428336819a2069c00"} Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.774843 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.823662 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5addc135-8f0f-4edd-ad29-76f3fd340827-crc-storage\") pod \"5addc135-8f0f-4edd-ad29-76f3fd340827\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.823736 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz7nt\" (UniqueName: \"kubernetes.io/projected/5addc135-8f0f-4edd-ad29-76f3fd340827-kube-api-access-vz7nt\") pod \"5addc135-8f0f-4edd-ad29-76f3fd340827\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.823936 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5addc135-8f0f-4edd-ad29-76f3fd340827-node-mnt\") pod \"5addc135-8f0f-4edd-ad29-76f3fd340827\" (UID: \"5addc135-8f0f-4edd-ad29-76f3fd340827\") " Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.823989 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5addc135-8f0f-4edd-ad29-76f3fd340827-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "5addc135-8f0f-4edd-ad29-76f3fd340827" (UID: "5addc135-8f0f-4edd-ad29-76f3fd340827"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.824281 4733 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5addc135-8f0f-4edd-ad29-76f3fd340827-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.832380 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5addc135-8f0f-4edd-ad29-76f3fd340827-kube-api-access-vz7nt" (OuterVolumeSpecName: "kube-api-access-vz7nt") pod "5addc135-8f0f-4edd-ad29-76f3fd340827" (UID: "5addc135-8f0f-4edd-ad29-76f3fd340827"). InnerVolumeSpecName "kube-api-access-vz7nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.851613 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5addc135-8f0f-4edd-ad29-76f3fd340827-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "5addc135-8f0f-4edd-ad29-76f3fd340827" (UID: "5addc135-8f0f-4edd-ad29-76f3fd340827"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.925635 4733 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5addc135-8f0f-4edd-ad29-76f3fd340827-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:29 crc kubenswrapper[4733]: I1204 17:52:29.926007 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz7nt\" (UniqueName: \"kubernetes.io/projected/5addc135-8f0f-4edd-ad29-76f3fd340827-kube-api-access-vz7nt\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:30 crc kubenswrapper[4733]: I1204 17:52:30.498693 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5jhj" Dec 04 17:52:30 crc kubenswrapper[4733]: I1204 17:52:30.498937 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-t5jhj" event={"ID":"5addc135-8f0f-4edd-ad29-76f3fd340827","Type":"ContainerDied","Data":"5ab60af3821462e94a340acfbff67263f71ec039fac233f01175612249f3c5db"} Dec 04 17:52:30 crc kubenswrapper[4733]: I1204 17:52:30.498992 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ab60af3821462e94a340acfbff67263f71ec039fac233f01175612249f3c5db" Dec 04 17:52:35 crc kubenswrapper[4733]: I1204 17:52:35.993768 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8p74n" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.528587 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6"] Dec 04 17:52:38 crc kubenswrapper[4733]: E1204 17:52:38.529219 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5addc135-8f0f-4edd-ad29-76f3fd340827" containerName="storage" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.529238 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5addc135-8f0f-4edd-ad29-76f3fd340827" containerName="storage" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.529408 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5addc135-8f0f-4edd-ad29-76f3fd340827" containerName="storage" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.530495 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.532603 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.543482 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6"] Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.670534 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ssmw\" (UniqueName: \"kubernetes.io/projected/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-kube-api-access-4ssmw\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.670600 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.670627 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.771629 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ssmw\" (UniqueName: \"kubernetes.io/projected/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-kube-api-access-4ssmw\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.771719 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.771754 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.772622 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.772928 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.800301 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ssmw\" (UniqueName: \"kubernetes.io/projected/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-kube-api-access-4ssmw\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:38 crc kubenswrapper[4733]: I1204 17:52:38.880312 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.135741 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6"] Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.586655 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" event={"ID":"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b","Type":"ContainerStarted","Data":"182a586dc7d016f50d7a51355afd68c0d8ea7e3da7deaf166b1b8d7e51404163"} Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.587220 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" event={"ID":"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b","Type":"ContainerStarted","Data":"7abcbdee27a8623b6a04244a6f83eee3df967a55420d721e6f0de2053af4c5ac"} Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.699938 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n2xgd"] Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.711301 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.727392 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n2xgd"] Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.794135 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-catalog-content\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.794201 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkszh\" (UniqueName: \"kubernetes.io/projected/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-kube-api-access-fkszh\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.794303 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-utilities\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.895336 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkszh\" (UniqueName: \"kubernetes.io/projected/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-kube-api-access-fkszh\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.895432 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-utilities\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.895500 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-catalog-content\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.896015 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-catalog-content\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.896075 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-utilities\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:39 crc kubenswrapper[4733]: I1204 17:52:39.921831 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkszh\" (UniqueName: \"kubernetes.io/projected/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-kube-api-access-fkszh\") pod \"redhat-operators-n2xgd\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:40 crc kubenswrapper[4733]: I1204 17:52:40.094683 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:40 crc kubenswrapper[4733]: I1204 17:52:40.351342 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n2xgd"] Dec 04 17:52:40 crc kubenswrapper[4733]: W1204 17:52:40.354328 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod228147e3_7ca5_4a9e_9aa1_3fbaa8ba562d.slice/crio-e2a8a7b83f58062ebbd71c3ab18b49ae12d48a022874dac507983a7df318c04c WatchSource:0}: Error finding container e2a8a7b83f58062ebbd71c3ab18b49ae12d48a022874dac507983a7df318c04c: Status 404 returned error can't find the container with id e2a8a7b83f58062ebbd71c3ab18b49ae12d48a022874dac507983a7df318c04c Dec 04 17:52:40 crc kubenswrapper[4733]: I1204 17:52:40.592093 4733 generic.go:334] "Generic (PLEG): container finished" podID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerID="7e579f8eb71f30ad8e869c173a8d3da368bb2323514cc5c598045a444f89cc5d" exitCode=0 Dec 04 17:52:40 crc kubenswrapper[4733]: I1204 17:52:40.592154 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2xgd" event={"ID":"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d","Type":"ContainerDied","Data":"7e579f8eb71f30ad8e869c173a8d3da368bb2323514cc5c598045a444f89cc5d"} Dec 04 17:52:40 crc kubenswrapper[4733]: I1204 17:52:40.592186 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2xgd" event={"ID":"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d","Type":"ContainerStarted","Data":"e2a8a7b83f58062ebbd71c3ab18b49ae12d48a022874dac507983a7df318c04c"} Dec 04 17:52:40 crc kubenswrapper[4733]: I1204 17:52:40.594633 4733 generic.go:334] "Generic (PLEG): container finished" podID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerID="182a586dc7d016f50d7a51355afd68c0d8ea7e3da7deaf166b1b8d7e51404163" exitCode=0 Dec 04 17:52:40 crc kubenswrapper[4733]: I1204 17:52:40.594686 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" event={"ID":"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b","Type":"ContainerDied","Data":"182a586dc7d016f50d7a51355afd68c0d8ea7e3da7deaf166b1b8d7e51404163"} Dec 04 17:52:41 crc kubenswrapper[4733]: I1204 17:52:41.601644 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2xgd" event={"ID":"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d","Type":"ContainerStarted","Data":"827f58f924af78c052d54ee4c987fcf62fe9ebc2841647c6dbd5255ef643cb81"} Dec 04 17:52:42 crc kubenswrapper[4733]: I1204 17:52:42.610845 4733 generic.go:334] "Generic (PLEG): container finished" podID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerID="e0256a7bd49fef7b22b138e7fd718f2955dd9584d26fdd7dd260e9acab4fdc07" exitCode=0 Dec 04 17:52:42 crc kubenswrapper[4733]: I1204 17:52:42.610994 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" event={"ID":"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b","Type":"ContainerDied","Data":"e0256a7bd49fef7b22b138e7fd718f2955dd9584d26fdd7dd260e9acab4fdc07"} Dec 04 17:52:42 crc kubenswrapper[4733]: I1204 17:52:42.614115 4733 generic.go:334] "Generic (PLEG): container finished" podID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerID="827f58f924af78c052d54ee4c987fcf62fe9ebc2841647c6dbd5255ef643cb81" exitCode=0 Dec 04 17:52:42 crc kubenswrapper[4733]: I1204 17:52:42.614179 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2xgd" event={"ID":"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d","Type":"ContainerDied","Data":"827f58f924af78c052d54ee4c987fcf62fe9ebc2841647c6dbd5255ef643cb81"} Dec 04 17:52:43 crc kubenswrapper[4733]: I1204 17:52:43.624529 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2xgd" event={"ID":"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d","Type":"ContainerStarted","Data":"3d08bd7143f42f1a15dac6025a291ad883bdc5d4ecd4f9a1390ae900f1ae4ad3"} Dec 04 17:52:43 crc kubenswrapper[4733]: I1204 17:52:43.628584 4733 generic.go:334] "Generic (PLEG): container finished" podID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerID="84a2d733f4800cbc572a1c870e881e4f28a3be137c86703fd8440a039eaa54b5" exitCode=0 Dec 04 17:52:43 crc kubenswrapper[4733]: I1204 17:52:43.628612 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" event={"ID":"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b","Type":"ContainerDied","Data":"84a2d733f4800cbc572a1c870e881e4f28a3be137c86703fd8440a039eaa54b5"} Dec 04 17:52:43 crc kubenswrapper[4733]: I1204 17:52:43.652977 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n2xgd" podStartSLOduration=2.206840111 podStartE2EDuration="4.65295485s" podCreationTimestamp="2025-12-04 17:52:39 +0000 UTC" firstStartedPulling="2025-12-04 17:52:40.593915936 +0000 UTC m=+822.549276972" lastFinishedPulling="2025-12-04 17:52:43.040030625 +0000 UTC m=+824.995391711" observedRunningTime="2025-12-04 17:52:43.649361091 +0000 UTC m=+825.604722207" watchObservedRunningTime="2025-12-04 17:52:43.65295485 +0000 UTC m=+825.608315906" Dec 04 17:52:44 crc kubenswrapper[4733]: I1204 17:52:44.934744 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.073271 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-bundle\") pod \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.073381 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ssmw\" (UniqueName: \"kubernetes.io/projected/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-kube-api-access-4ssmw\") pod \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.073505 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-util\") pod \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\" (UID: \"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b\") " Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.074137 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-bundle" (OuterVolumeSpecName: "bundle") pod "4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" (UID: "4209b6af-a01b-4f51-8cf7-d5bc796c4a3b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.079461 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-kube-api-access-4ssmw" (OuterVolumeSpecName: "kube-api-access-4ssmw") pod "4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" (UID: "4209b6af-a01b-4f51-8cf7-d5bc796c4a3b"). InnerVolumeSpecName "kube-api-access-4ssmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.088070 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-util" (OuterVolumeSpecName: "util") pod "4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" (UID: "4209b6af-a01b-4f51-8cf7-d5bc796c4a3b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.175077 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-util\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.175138 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.175151 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ssmw\" (UniqueName: \"kubernetes.io/projected/4209b6af-a01b-4f51-8cf7-d5bc796c4a3b-kube-api-access-4ssmw\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.362199 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.362287 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.362356 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.363192 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee5a68aff544c2577446d346027585de3353ade4bb096661850d64dabfc79180"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.363322 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://ee5a68aff544c2577446d346027585de3353ade4bb096661850d64dabfc79180" gracePeriod=600 Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.643336 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="ee5a68aff544c2577446d346027585de3353ade4bb096661850d64dabfc79180" exitCode=0 Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.643423 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"ee5a68aff544c2577446d346027585de3353ade4bb096661850d64dabfc79180"} Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.643504 4733 scope.go:117] "RemoveContainer" containerID="d83e5743bfe34b630102892e9bcb5c583d801852811e612149b74d0af61af29c" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.646817 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" event={"ID":"4209b6af-a01b-4f51-8cf7-d5bc796c4a3b","Type":"ContainerDied","Data":"7abcbdee27a8623b6a04244a6f83eee3df967a55420d721e6f0de2053af4c5ac"} Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.646851 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7abcbdee27a8623b6a04244a6f83eee3df967a55420d721e6f0de2053af4c5ac" Dec 04 17:52:45 crc kubenswrapper[4733]: I1204 17:52:45.646900 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6" Dec 04 17:52:46 crc kubenswrapper[4733]: I1204 17:52:46.652714 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"245d5f7c913eadd0235bcdf1e7f003bb02e52afa3a7ef2fd2c9a9a3603961fd8"} Dec 04 17:52:46 crc kubenswrapper[4733]: I1204 17:52:46.992756 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7"] Dec 04 17:52:46 crc kubenswrapper[4733]: E1204 17:52:46.992989 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerName="pull" Dec 04 17:52:46 crc kubenswrapper[4733]: I1204 17:52:46.993002 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerName="pull" Dec 04 17:52:46 crc kubenswrapper[4733]: E1204 17:52:46.993011 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerName="extract" Dec 04 17:52:46 crc kubenswrapper[4733]: I1204 17:52:46.993017 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerName="extract" Dec 04 17:52:46 crc kubenswrapper[4733]: E1204 17:52:46.993030 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerName="util" Dec 04 17:52:46 crc kubenswrapper[4733]: I1204 17:52:46.993036 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerName="util" Dec 04 17:52:46 crc kubenswrapper[4733]: I1204 17:52:46.993143 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4209b6af-a01b-4f51-8cf7-d5bc796c4a3b" containerName="extract" Dec 04 17:52:46 crc kubenswrapper[4733]: I1204 17:52:46.993569 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7" Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.000069 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-lz7fv" Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.000159 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.000175 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.012728 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7"] Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.101669 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b6gt\" (UniqueName: \"kubernetes.io/projected/57ea0cda-c1a7-400c-89b5-350c5f6a1bb7-kube-api-access-5b6gt\") pod \"nmstate-operator-5b5b58f5c8-7rgd7\" (UID: \"57ea0cda-c1a7-400c-89b5-350c5f6a1bb7\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7" Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.202584 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b6gt\" (UniqueName: \"kubernetes.io/projected/57ea0cda-c1a7-400c-89b5-350c5f6a1bb7-kube-api-access-5b6gt\") pod \"nmstate-operator-5b5b58f5c8-7rgd7\" (UID: \"57ea0cda-c1a7-400c-89b5-350c5f6a1bb7\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7" Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.220167 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b6gt\" (UniqueName: \"kubernetes.io/projected/57ea0cda-c1a7-400c-89b5-350c5f6a1bb7-kube-api-access-5b6gt\") pod \"nmstate-operator-5b5b58f5c8-7rgd7\" (UID: \"57ea0cda-c1a7-400c-89b5-350c5f6a1bb7\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7" Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.308465 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7" Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.514848 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7"] Dec 04 17:52:47 crc kubenswrapper[4733]: W1204 17:52:47.519873 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57ea0cda_c1a7_400c_89b5_350c5f6a1bb7.slice/crio-651e668aec4d2354353773379b911ce1a2a5189f52c2581f517e63ebbcb03cda WatchSource:0}: Error finding container 651e668aec4d2354353773379b911ce1a2a5189f52c2581f517e63ebbcb03cda: Status 404 returned error can't find the container with id 651e668aec4d2354353773379b911ce1a2a5189f52c2581f517e63ebbcb03cda Dec 04 17:52:47 crc kubenswrapper[4733]: I1204 17:52:47.662655 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7" event={"ID":"57ea0cda-c1a7-400c-89b5-350c5f6a1bb7","Type":"ContainerStarted","Data":"651e668aec4d2354353773379b911ce1a2a5189f52c2581f517e63ebbcb03cda"} Dec 04 17:52:50 crc kubenswrapper[4733]: I1204 17:52:50.095201 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:50 crc kubenswrapper[4733]: I1204 17:52:50.095775 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:50 crc kubenswrapper[4733]: I1204 17:52:50.157510 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:50 crc kubenswrapper[4733]: I1204 17:52:50.733283 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:51 crc kubenswrapper[4733]: I1204 17:52:51.690144 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7" event={"ID":"57ea0cda-c1a7-400c-89b5-350c5f6a1bb7","Type":"ContainerStarted","Data":"bacd9a75e2ccac13f0675c7f867f928b4cca5b6cba6d9721cd9db5299656859f"} Dec 04 17:52:51 crc kubenswrapper[4733]: I1204 17:52:51.721723 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7rgd7" podStartSLOduration=2.4881353219999998 podStartE2EDuration="5.721701213s" podCreationTimestamp="2025-12-04 17:52:46 +0000 UTC" firstStartedPulling="2025-12-04 17:52:47.521791445 +0000 UTC m=+829.477152501" lastFinishedPulling="2025-12-04 17:52:50.755357336 +0000 UTC m=+832.710718392" observedRunningTime="2025-12-04 17:52:51.714324812 +0000 UTC m=+833.669685868" watchObservedRunningTime="2025-12-04 17:52:51.721701213 +0000 UTC m=+833.677062269" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.479045 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n2xgd"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.603900 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.604932 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.609841 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-lb4tq" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.614332 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.615309 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.616835 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.648440 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-l2pgf"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.649548 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.677541 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn6cv\" (UniqueName: \"kubernetes.io/projected/3c0e12d4-71fa-444b-a3e7-b47c798053f5-kube-api-access-qn6cv\") pod \"nmstate-metrics-7f946cbc9-grw9m\" (UID: \"3c0e12d4-71fa-444b-a3e7-b47c798053f5\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.679697 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.689097 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.695177 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n2xgd" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerName="registry-server" containerID="cri-o://3d08bd7143f42f1a15dac6025a291ad883bdc5d4ecd4f9a1390ae900f1ae4ad3" gracePeriod=2 Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.728902 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.729528 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.732145 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.733473 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-rk8pj" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.738215 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.747186 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.778872 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-dbus-socket\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.778907 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl57m\" (UniqueName: \"kubernetes.io/projected/cbba5431-8bfc-4fe8-b496-29993e49d748-kube-api-access-zl57m\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.779062 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a5af3730-230a-4686-8262-bef9e42c60b5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-v2whn\" (UID: \"a5af3730-230a-4686-8262-bef9e42c60b5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.779119 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6p4b\" (UniqueName: \"kubernetes.io/projected/a5af3730-230a-4686-8262-bef9e42c60b5-kube-api-access-t6p4b\") pod \"nmstate-webhook-5f6d4c5ccb-v2whn\" (UID: \"a5af3730-230a-4686-8262-bef9e42c60b5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.779278 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn6cv\" (UniqueName: \"kubernetes.io/projected/3c0e12d4-71fa-444b-a3e7-b47c798053f5-kube-api-access-qn6cv\") pod \"nmstate-metrics-7f946cbc9-grw9m\" (UID: \"3c0e12d4-71fa-444b-a3e7-b47c798053f5\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.779366 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-ovs-socket\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.779411 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-nmstate-lock\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.799753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn6cv\" (UniqueName: \"kubernetes.io/projected/3c0e12d4-71fa-444b-a3e7-b47c798053f5-kube-api-access-qn6cv\") pod \"nmstate-metrics-7f946cbc9-grw9m\" (UID: \"3c0e12d4-71fa-444b-a3e7-b47c798053f5\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881021 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-ovs-socket\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881077 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-nmstate-lock\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881128 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881155 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-dbus-socket\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881178 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl57m\" (UniqueName: \"kubernetes.io/projected/cbba5431-8bfc-4fe8-b496-29993e49d748-kube-api-access-zl57m\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881206 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881229 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6ncw\" (UniqueName: \"kubernetes.io/projected/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-kube-api-access-z6ncw\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881263 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a5af3730-230a-4686-8262-bef9e42c60b5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-v2whn\" (UID: \"a5af3730-230a-4686-8262-bef9e42c60b5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881286 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6p4b\" (UniqueName: \"kubernetes.io/projected/a5af3730-230a-4686-8262-bef9e42c60b5-kube-api-access-t6p4b\") pod \"nmstate-webhook-5f6d4c5ccb-v2whn\" (UID: \"a5af3730-230a-4686-8262-bef9e42c60b5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881611 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-ovs-socket\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881660 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-nmstate-lock\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.881945 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cbba5431-8bfc-4fe8-b496-29993e49d748-dbus-socket\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: E1204 17:52:52.882049 4733 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 04 17:52:52 crc kubenswrapper[4733]: E1204 17:52:52.882107 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a5af3730-230a-4686-8262-bef9e42c60b5-tls-key-pair podName:a5af3730-230a-4686-8262-bef9e42c60b5 nodeName:}" failed. No retries permitted until 2025-12-04 17:52:53.3820883 +0000 UTC m=+835.337449346 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/a5af3730-230a-4686-8262-bef9e42c60b5-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-v2whn" (UID: "a5af3730-230a-4686-8262-bef9e42c60b5") : secret "openshift-nmstate-webhook" not found Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.910431 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl57m\" (UniqueName: \"kubernetes.io/projected/cbba5431-8bfc-4fe8-b496-29993e49d748-kube-api-access-zl57m\") pod \"nmstate-handler-l2pgf\" (UID: \"cbba5431-8bfc-4fe8-b496-29993e49d748\") " pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.910623 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6p4b\" (UniqueName: \"kubernetes.io/projected/a5af3730-230a-4686-8262-bef9e42c60b5-kube-api-access-t6p4b\") pod \"nmstate-webhook-5f6d4c5ccb-v2whn\" (UID: \"a5af3730-230a-4686-8262-bef9e42c60b5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.927332 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-75f75bdd6-fl625"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.928218 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.932332 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75f75bdd6-fl625"] Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.948702 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.970195 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.983317 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.983367 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.983385 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6ncw\" (UniqueName: \"kubernetes.io/projected/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-kube-api-access-z6ncw\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.984449 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.987384 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:52 crc kubenswrapper[4733]: W1204 17:52:52.991063 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbba5431_8bfc_4fe8_b496_29993e49d748.slice/crio-78ba7abfa5e3ceb1283870318f375dbf1f58e351de0483d68e184c2b904b6504 WatchSource:0}: Error finding container 78ba7abfa5e3ceb1283870318f375dbf1f58e351de0483d68e184c2b904b6504: Status 404 returned error can't find the container with id 78ba7abfa5e3ceb1283870318f375dbf1f58e351de0483d68e184c2b904b6504 Dec 04 17:52:52 crc kubenswrapper[4733]: I1204 17:52:52.999190 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6ncw\" (UniqueName: \"kubernetes.io/projected/ce10fe7f-07b2-4f30-a52c-1fc47a486ff3-kube-api-access-z6ncw\") pod \"nmstate-console-plugin-7fbb5f6569-z84kw\" (UID: \"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.053855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.084353 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-oauth-serving-cert\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.084402 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-service-ca\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.084428 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-console-config\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.084446 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-trusted-ca-bundle\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.084652 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4831aa7-97f5-4097-9d82-3202f855e65c-console-serving-cert\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.084807 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4831aa7-97f5-4097-9d82-3202f855e65c-console-oauth-config\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.084850 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dx82\" (UniqueName: \"kubernetes.io/projected/e4831aa7-97f5-4097-9d82-3202f855e65c-kube-api-access-4dx82\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.140878 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m"] Dec 04 17:52:53 crc kubenswrapper[4733]: W1204 17:52:53.148212 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c0e12d4_71fa_444b_a3e7_b47c798053f5.slice/crio-a8f3a9ff1f44ed539b7b133276feaf64ac8c92cf53b8ee1f7397e096bba80188 WatchSource:0}: Error finding container a8f3a9ff1f44ed539b7b133276feaf64ac8c92cf53b8ee1f7397e096bba80188: Status 404 returned error can't find the container with id a8f3a9ff1f44ed539b7b133276feaf64ac8c92cf53b8ee1f7397e096bba80188 Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.186179 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-oauth-serving-cert\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.186226 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-service-ca\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.186265 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-console-config\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.186285 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-trusted-ca-bundle\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.186332 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4831aa7-97f5-4097-9d82-3202f855e65c-console-serving-cert\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.186386 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4831aa7-97f5-4097-9d82-3202f855e65c-console-oauth-config\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.186409 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dx82\" (UniqueName: \"kubernetes.io/projected/e4831aa7-97f5-4097-9d82-3202f855e65c-kube-api-access-4dx82\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.187028 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-oauth-serving-cert\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.187857 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-console-config\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.187885 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-service-ca\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.188022 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4831aa7-97f5-4097-9d82-3202f855e65c-trusted-ca-bundle\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.193469 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4831aa7-97f5-4097-9d82-3202f855e65c-console-oauth-config\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.194522 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4831aa7-97f5-4097-9d82-3202f855e65c-console-serving-cert\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.205467 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dx82\" (UniqueName: \"kubernetes.io/projected/e4831aa7-97f5-4097-9d82-3202f855e65c-kube-api-access-4dx82\") pod \"console-75f75bdd6-fl625\" (UID: \"e4831aa7-97f5-4097-9d82-3202f855e65c\") " pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.231672 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw"] Dec 04 17:52:53 crc kubenswrapper[4733]: W1204 17:52:53.235549 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce10fe7f_07b2_4f30_a52c_1fc47a486ff3.slice/crio-48951fed9f7d3a21d38478f7b6ac1c8e9fd29d58e3c37e755ce4412cecd4d71a WatchSource:0}: Error finding container 48951fed9f7d3a21d38478f7b6ac1c8e9fd29d58e3c37e755ce4412cecd4d71a: Status 404 returned error can't find the container with id 48951fed9f7d3a21d38478f7b6ac1c8e9fd29d58e3c37e755ce4412cecd4d71a Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.259820 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.388454 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a5af3730-230a-4686-8262-bef9e42c60b5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-v2whn\" (UID: \"a5af3730-230a-4686-8262-bef9e42c60b5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.393497 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a5af3730-230a-4686-8262-bef9e42c60b5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-v2whn\" (UID: \"a5af3730-230a-4686-8262-bef9e42c60b5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.454987 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75f75bdd6-fl625"] Dec 04 17:52:53 crc kubenswrapper[4733]: W1204 17:52:53.461355 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4831aa7_97f5_4097_9d82_3202f855e65c.slice/crio-af37d095606cf50233fbda7084f935d19aa3573cd8ba1a476e24b71e5bf9a28b WatchSource:0}: Error finding container af37d095606cf50233fbda7084f935d19aa3573cd8ba1a476e24b71e5bf9a28b: Status 404 returned error can't find the container with id af37d095606cf50233fbda7084f935d19aa3573cd8ba1a476e24b71e5bf9a28b Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.560883 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.703691 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-l2pgf" event={"ID":"cbba5431-8bfc-4fe8-b496-29993e49d748","Type":"ContainerStarted","Data":"78ba7abfa5e3ceb1283870318f375dbf1f58e351de0483d68e184c2b904b6504"} Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.705233 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" event={"ID":"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3","Type":"ContainerStarted","Data":"48951fed9f7d3a21d38478f7b6ac1c8e9fd29d58e3c37e755ce4412cecd4d71a"} Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.706379 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" event={"ID":"3c0e12d4-71fa-444b-a3e7-b47c798053f5","Type":"ContainerStarted","Data":"a8f3a9ff1f44ed539b7b133276feaf64ac8c92cf53b8ee1f7397e096bba80188"} Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.708057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75f75bdd6-fl625" event={"ID":"e4831aa7-97f5-4097-9d82-3202f855e65c","Type":"ContainerStarted","Data":"af37d095606cf50233fbda7084f935d19aa3573cd8ba1a476e24b71e5bf9a28b"} Dec 04 17:52:53 crc kubenswrapper[4733]: W1204 17:52:53.785270 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5af3730_230a_4686_8262_bef9e42c60b5.slice/crio-aa9ef783ccd088abc0ce098b205d2475f7660d85ecfbf6d9749f3df5cfa5fa3d WatchSource:0}: Error finding container aa9ef783ccd088abc0ce098b205d2475f7660d85ecfbf6d9749f3df5cfa5fa3d: Status 404 returned error can't find the container with id aa9ef783ccd088abc0ce098b205d2475f7660d85ecfbf6d9749f3df5cfa5fa3d Dec 04 17:52:53 crc kubenswrapper[4733]: I1204 17:52:53.786026 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn"] Dec 04 17:52:54 crc kubenswrapper[4733]: I1204 17:52:54.718417 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" event={"ID":"a5af3730-230a-4686-8262-bef9e42c60b5","Type":"ContainerStarted","Data":"aa9ef783ccd088abc0ce098b205d2475f7660d85ecfbf6d9749f3df5cfa5fa3d"} Dec 04 17:52:55 crc kubenswrapper[4733]: I1204 17:52:55.725168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75f75bdd6-fl625" event={"ID":"e4831aa7-97f5-4097-9d82-3202f855e65c","Type":"ContainerStarted","Data":"ca178518786d4c3aeb44027752d18c4220ddffc02651f9ce2aa160598d3a0227"} Dec 04 17:52:55 crc kubenswrapper[4733]: I1204 17:52:55.729827 4733 generic.go:334] "Generic (PLEG): container finished" podID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerID="3d08bd7143f42f1a15dac6025a291ad883bdc5d4ecd4f9a1390ae900f1ae4ad3" exitCode=0 Dec 04 17:52:55 crc kubenswrapper[4733]: I1204 17:52:55.729866 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2xgd" event={"ID":"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d","Type":"ContainerDied","Data":"3d08bd7143f42f1a15dac6025a291ad883bdc5d4ecd4f9a1390ae900f1ae4ad3"} Dec 04 17:52:55 crc kubenswrapper[4733]: I1204 17:52:55.750346 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-75f75bdd6-fl625" podStartSLOduration=3.750302273 podStartE2EDuration="3.750302273s" podCreationTimestamp="2025-12-04 17:52:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:52:55.750295753 +0000 UTC m=+837.705656799" watchObservedRunningTime="2025-12-04 17:52:55.750302273 +0000 UTC m=+837.705663319" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.032924 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.131272 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-utilities\") pod \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.131332 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-catalog-content\") pod \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.131361 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkszh\" (UniqueName: \"kubernetes.io/projected/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-kube-api-access-fkszh\") pod \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\" (UID: \"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d\") " Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.133436 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-utilities" (OuterVolumeSpecName: "utilities") pod "228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" (UID: "228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.143112 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-kube-api-access-fkszh" (OuterVolumeSpecName: "kube-api-access-fkszh") pod "228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" (UID: "228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d"). InnerVolumeSpecName "kube-api-access-fkszh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.232952 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.232991 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkszh\" (UniqueName: \"kubernetes.io/projected/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-kube-api-access-fkszh\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.239444 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" (UID: "228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.334324 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.737423 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2xgd" event={"ID":"228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d","Type":"ContainerDied","Data":"e2a8a7b83f58062ebbd71c3ab18b49ae12d48a022874dac507983a7df318c04c"} Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.737511 4733 scope.go:117] "RemoveContainer" containerID="3d08bd7143f42f1a15dac6025a291ad883bdc5d4ecd4f9a1390ae900f1ae4ad3" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.737463 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n2xgd" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.756983 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n2xgd"] Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.760935 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n2xgd"] Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.778615 4733 scope.go:117] "RemoveContainer" containerID="827f58f924af78c052d54ee4c987fcf62fe9ebc2841647c6dbd5255ef643cb81" Dec 04 17:52:56 crc kubenswrapper[4733]: I1204 17:52:56.828029 4733 scope.go:117] "RemoveContainer" containerID="7e579f8eb71f30ad8e869c173a8d3da368bb2323514cc5c598045a444f89cc5d" Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.750549 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-l2pgf" event={"ID":"cbba5431-8bfc-4fe8-b496-29993e49d748","Type":"ContainerStarted","Data":"7ae946f951ebdb58cab47476b09ab385ee398640879ca8f97a56b9708699e85e"} Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.751345 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.754224 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" event={"ID":"a5af3730-230a-4686-8262-bef9e42c60b5","Type":"ContainerStarted","Data":"8b6480274cbf6d7e3e63a63fcfd20d8c6175d84f521754bff5b90b3db8ea8840"} Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.754339 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.760258 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" event={"ID":"ce10fe7f-07b2-4f30-a52c-1fc47a486ff3","Type":"ContainerStarted","Data":"1c341af4d3bf1b0ada764a38b0691e966a0d51c9431601a97fb0af4b160bd986"} Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.763083 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" event={"ID":"3c0e12d4-71fa-444b-a3e7-b47c798053f5","Type":"ContainerStarted","Data":"e06430e0910f28546c0d44ff98bb07691db922f4f8ce6bb7144ff39a62cc4813"} Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.782756 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-l2pgf" podStartSLOduration=1.991954299 podStartE2EDuration="5.782727247s" podCreationTimestamp="2025-12-04 17:52:52 +0000 UTC" firstStartedPulling="2025-12-04 17:52:52.993586869 +0000 UTC m=+834.948947915" lastFinishedPulling="2025-12-04 17:52:56.784359807 +0000 UTC m=+838.739720863" observedRunningTime="2025-12-04 17:52:57.777621048 +0000 UTC m=+839.732982134" watchObservedRunningTime="2025-12-04 17:52:57.782727247 +0000 UTC m=+839.738088303" Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.811269 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-z84kw" podStartSLOduration=2.28326366 podStartE2EDuration="5.811239455s" podCreationTimestamp="2025-12-04 17:52:52 +0000 UTC" firstStartedPulling="2025-12-04 17:52:53.23761361 +0000 UTC m=+835.192974656" lastFinishedPulling="2025-12-04 17:52:56.765589405 +0000 UTC m=+838.720950451" observedRunningTime="2025-12-04 17:52:57.802057424 +0000 UTC m=+839.757418540" watchObservedRunningTime="2025-12-04 17:52:57.811239455 +0000 UTC m=+839.766600541" Dec 04 17:52:57 crc kubenswrapper[4733]: I1204 17:52:57.835526 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" podStartSLOduration=2.834517862 podStartE2EDuration="5.835489185s" podCreationTimestamp="2025-12-04 17:52:52 +0000 UTC" firstStartedPulling="2025-12-04 17:52:53.787672501 +0000 UTC m=+835.743033557" lastFinishedPulling="2025-12-04 17:52:56.788643824 +0000 UTC m=+838.744004880" observedRunningTime="2025-12-04 17:52:57.832345919 +0000 UTC m=+839.787706975" watchObservedRunningTime="2025-12-04 17:52:57.835489185 +0000 UTC m=+839.790850271" Dec 04 17:52:58 crc kubenswrapper[4733]: I1204 17:52:58.348296 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" path="/var/lib/kubelet/pods/228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d/volumes" Dec 04 17:52:58 crc kubenswrapper[4733]: I1204 17:52:58.711095 4733 scope.go:117] "RemoveContainer" containerID="dffee5a0cc39606fc396c202bec1e4d1069b46f3d5c83da324c0adf333957886" Dec 04 17:52:58 crc kubenswrapper[4733]: I1204 17:52:58.771080 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-472m8_829fda7a-6a96-4007-b2f5-0be03782a764/kube-multus/2.log" Dec 04 17:52:59 crc kubenswrapper[4733]: I1204 17:52:59.779304 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" event={"ID":"3c0e12d4-71fa-444b-a3e7-b47c798053f5","Type":"ContainerStarted","Data":"0d0c7ec9b2e2bc8f37df669cee313261b15ba994da547da2377025e45e83c8d5"} Dec 04 17:52:59 crc kubenswrapper[4733]: I1204 17:52:59.808175 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-grw9m" podStartSLOduration=1.572105585 podStartE2EDuration="7.808138239s" podCreationTimestamp="2025-12-04 17:52:52 +0000 UTC" firstStartedPulling="2025-12-04 17:52:53.151008309 +0000 UTC m=+835.106369355" lastFinishedPulling="2025-12-04 17:52:59.387040963 +0000 UTC m=+841.342402009" observedRunningTime="2025-12-04 17:52:59.803777941 +0000 UTC m=+841.759139027" watchObservedRunningTime="2025-12-04 17:52:59.808138239 +0000 UTC m=+841.763499325" Dec 04 17:53:03 crc kubenswrapper[4733]: I1204 17:53:03.006887 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-l2pgf" Dec 04 17:53:03 crc kubenswrapper[4733]: I1204 17:53:03.260840 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:53:03 crc kubenswrapper[4733]: I1204 17:53:03.260943 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:53:03 crc kubenswrapper[4733]: I1204 17:53:03.268916 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:53:03 crc kubenswrapper[4733]: I1204 17:53:03.810268 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-75f75bdd6-fl625" Dec 04 17:53:03 crc kubenswrapper[4733]: I1204 17:53:03.891680 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rkdfs"] Dec 04 17:53:13 crc kubenswrapper[4733]: I1204 17:53:13.569075 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-v2whn" Dec 04 17:53:28 crc kubenswrapper[4733]: I1204 17:53:28.948889 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-rkdfs" podUID="6c99fe8f-586c-4e35-b010-73dad59c59af" containerName="console" containerID="cri-o://2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc" gracePeriod=15 Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.326835 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rkdfs_6c99fe8f-586c-4e35-b010-73dad59c59af/console/0.log" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.327151 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.434228 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6q7d\" (UniqueName: \"kubernetes.io/projected/6c99fe8f-586c-4e35-b010-73dad59c59af-kube-api-access-j6q7d\") pod \"6c99fe8f-586c-4e35-b010-73dad59c59af\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.434305 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-oauth-serving-cert\") pod \"6c99fe8f-586c-4e35-b010-73dad59c59af\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.434383 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-serving-cert\") pod \"6c99fe8f-586c-4e35-b010-73dad59c59af\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.434431 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-oauth-config\") pod \"6c99fe8f-586c-4e35-b010-73dad59c59af\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.434474 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-console-config\") pod \"6c99fe8f-586c-4e35-b010-73dad59c59af\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.434544 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-service-ca\") pod \"6c99fe8f-586c-4e35-b010-73dad59c59af\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.434603 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-trusted-ca-bundle\") pod \"6c99fe8f-586c-4e35-b010-73dad59c59af\" (UID: \"6c99fe8f-586c-4e35-b010-73dad59c59af\") " Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.435284 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6c99fe8f-586c-4e35-b010-73dad59c59af" (UID: "6c99fe8f-586c-4e35-b010-73dad59c59af"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.435786 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-console-config" (OuterVolumeSpecName: "console-config") pod "6c99fe8f-586c-4e35-b010-73dad59c59af" (UID: "6c99fe8f-586c-4e35-b010-73dad59c59af"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.436350 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-service-ca" (OuterVolumeSpecName: "service-ca") pod "6c99fe8f-586c-4e35-b010-73dad59c59af" (UID: "6c99fe8f-586c-4e35-b010-73dad59c59af"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.436694 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6c99fe8f-586c-4e35-b010-73dad59c59af" (UID: "6c99fe8f-586c-4e35-b010-73dad59c59af"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.441546 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c99fe8f-586c-4e35-b010-73dad59c59af-kube-api-access-j6q7d" (OuterVolumeSpecName: "kube-api-access-j6q7d") pod "6c99fe8f-586c-4e35-b010-73dad59c59af" (UID: "6c99fe8f-586c-4e35-b010-73dad59c59af"). InnerVolumeSpecName "kube-api-access-j6q7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.447998 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6c99fe8f-586c-4e35-b010-73dad59c59af" (UID: "6c99fe8f-586c-4e35-b010-73dad59c59af"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.448303 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6c99fe8f-586c-4e35-b010-73dad59c59af" (UID: "6c99fe8f-586c-4e35-b010-73dad59c59af"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.536460 4733 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.536513 4733 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.536533 4733 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6c99fe8f-586c-4e35-b010-73dad59c59af-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.536550 4733 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.536567 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.536585 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c99fe8f-586c-4e35-b010-73dad59c59af-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.536603 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6q7d\" (UniqueName: \"kubernetes.io/projected/6c99fe8f-586c-4e35-b010-73dad59c59af-kube-api-access-j6q7d\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.994847 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rkdfs_6c99fe8f-586c-4e35-b010-73dad59c59af/console/0.log" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.995096 4733 generic.go:334] "Generic (PLEG): container finished" podID="6c99fe8f-586c-4e35-b010-73dad59c59af" containerID="2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc" exitCode=2 Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.995123 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rkdfs" event={"ID":"6c99fe8f-586c-4e35-b010-73dad59c59af","Type":"ContainerDied","Data":"2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc"} Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.995149 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rkdfs" event={"ID":"6c99fe8f-586c-4e35-b010-73dad59c59af","Type":"ContainerDied","Data":"d945ec397b06f1c9f8b95b2d720c7ec23c87b5a85357c8d8fef53ea99fcd8d35"} Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.995166 4733 scope.go:117] "RemoveContainer" containerID="2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc" Dec 04 17:53:29 crc kubenswrapper[4733]: I1204 17:53:29.995217 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rkdfs" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.035465 4733 scope.go:117] "RemoveContainer" containerID="2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc" Dec 04 17:53:30 crc kubenswrapper[4733]: E1204 17:53:30.048862 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc\": container with ID starting with 2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc not found: ID does not exist" containerID="2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.048932 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc"} err="failed to get container status \"2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc\": rpc error: code = NotFound desc = could not find container \"2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc\": container with ID starting with 2f5396ecb9482167875f229ac5a6e8a98ce0e1f1a9eab0d79ae8c6b13d69abdc not found: ID does not exist" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.062483 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rkdfs"] Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.076168 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-rkdfs"] Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.347780 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c99fe8f-586c-4e35-b010-73dad59c59af" path="/var/lib/kubelet/pods/6c99fe8f-586c-4e35-b010-73dad59c59af/volumes" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.468238 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct"] Dec 04 17:53:30 crc kubenswrapper[4733]: E1204 17:53:30.468603 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerName="extract-content" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.468626 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerName="extract-content" Dec 04 17:53:30 crc kubenswrapper[4733]: E1204 17:53:30.468646 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerName="extract-utilities" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.468658 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerName="extract-utilities" Dec 04 17:53:30 crc kubenswrapper[4733]: E1204 17:53:30.468682 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerName="registry-server" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.468695 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerName="registry-server" Dec 04 17:53:30 crc kubenswrapper[4733]: E1204 17:53:30.468711 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c99fe8f-586c-4e35-b010-73dad59c59af" containerName="console" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.468723 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c99fe8f-586c-4e35-b010-73dad59c59af" containerName="console" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.468928 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="228147e3-7ca5-4a9e-9aa1-3fbaa8ba562d" containerName="registry-server" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.468947 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c99fe8f-586c-4e35-b010-73dad59c59af" containerName="console" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.470197 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.474137 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.482078 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct"] Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.557384 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.557432 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.557471 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6l24\" (UniqueName: \"kubernetes.io/projected/d54cd1df-e504-4d88-8d0d-41610c92c294-kube-api-access-r6l24\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.658676 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.658736 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.658786 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6l24\" (UniqueName: \"kubernetes.io/projected/d54cd1df-e504-4d88-8d0d-41610c92c294-kube-api-access-r6l24\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.659479 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.659776 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.699305 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6l24\" (UniqueName: \"kubernetes.io/projected/d54cd1df-e504-4d88-8d0d-41610c92c294-kube-api-access-r6l24\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:30 crc kubenswrapper[4733]: I1204 17:53:30.800825 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:31 crc kubenswrapper[4733]: I1204 17:53:31.073746 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct"] Dec 04 17:53:32 crc kubenswrapper[4733]: I1204 17:53:32.013270 4733 generic.go:334] "Generic (PLEG): container finished" podID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerID="7db9327582a38dcd1260477cf42d818617220a440bcae6ff240d70b52f1f4c5d" exitCode=0 Dec 04 17:53:32 crc kubenswrapper[4733]: I1204 17:53:32.013345 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" event={"ID":"d54cd1df-e504-4d88-8d0d-41610c92c294","Type":"ContainerDied","Data":"7db9327582a38dcd1260477cf42d818617220a440bcae6ff240d70b52f1f4c5d"} Dec 04 17:53:32 crc kubenswrapper[4733]: I1204 17:53:32.013539 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" event={"ID":"d54cd1df-e504-4d88-8d0d-41610c92c294","Type":"ContainerStarted","Data":"3ec5f4dab1b4de97e83e8ad8ac525bac9abb9f296786376fa7dd8a9b4ac159b3"} Dec 04 17:53:34 crc kubenswrapper[4733]: I1204 17:53:34.035351 4733 generic.go:334] "Generic (PLEG): container finished" podID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerID="821a604260ebee6579c30cf0ff7751830fcb9a7f0d29569b1cd58beca6e617ba" exitCode=0 Dec 04 17:53:34 crc kubenswrapper[4733]: I1204 17:53:34.035424 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" event={"ID":"d54cd1df-e504-4d88-8d0d-41610c92c294","Type":"ContainerDied","Data":"821a604260ebee6579c30cf0ff7751830fcb9a7f0d29569b1cd58beca6e617ba"} Dec 04 17:53:35 crc kubenswrapper[4733]: I1204 17:53:35.048230 4733 generic.go:334] "Generic (PLEG): container finished" podID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerID="dafcd8be7eb94656cd71d290ffbb7a0029313c1762e77a29cdec0f0dfa1d3be3" exitCode=0 Dec 04 17:53:35 crc kubenswrapper[4733]: I1204 17:53:35.048371 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" event={"ID":"d54cd1df-e504-4d88-8d0d-41610c92c294","Type":"ContainerDied","Data":"dafcd8be7eb94656cd71d290ffbb7a0029313c1762e77a29cdec0f0dfa1d3be3"} Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.380316 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.544242 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6l24\" (UniqueName: \"kubernetes.io/projected/d54cd1df-e504-4d88-8d0d-41610c92c294-kube-api-access-r6l24\") pod \"d54cd1df-e504-4d88-8d0d-41610c92c294\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.544318 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-util\") pod \"d54cd1df-e504-4d88-8d0d-41610c92c294\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.544347 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-bundle\") pod \"d54cd1df-e504-4d88-8d0d-41610c92c294\" (UID: \"d54cd1df-e504-4d88-8d0d-41610c92c294\") " Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.545427 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-bundle" (OuterVolumeSpecName: "bundle") pod "d54cd1df-e504-4d88-8d0d-41610c92c294" (UID: "d54cd1df-e504-4d88-8d0d-41610c92c294"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.553692 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d54cd1df-e504-4d88-8d0d-41610c92c294-kube-api-access-r6l24" (OuterVolumeSpecName: "kube-api-access-r6l24") pod "d54cd1df-e504-4d88-8d0d-41610c92c294" (UID: "d54cd1df-e504-4d88-8d0d-41610c92c294"). InnerVolumeSpecName "kube-api-access-r6l24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.645914 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.645983 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6l24\" (UniqueName: \"kubernetes.io/projected/d54cd1df-e504-4d88-8d0d-41610c92c294-kube-api-access-r6l24\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.798274 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-util" (OuterVolumeSpecName: "util") pod "d54cd1df-e504-4d88-8d0d-41610c92c294" (UID: "d54cd1df-e504-4d88-8d0d-41610c92c294"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:53:36 crc kubenswrapper[4733]: I1204 17:53:36.848755 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54cd1df-e504-4d88-8d0d-41610c92c294-util\") on node \"crc\" DevicePath \"\"" Dec 04 17:53:37 crc kubenswrapper[4733]: I1204 17:53:37.067704 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" event={"ID":"d54cd1df-e504-4d88-8d0d-41610c92c294","Type":"ContainerDied","Data":"3ec5f4dab1b4de97e83e8ad8ac525bac9abb9f296786376fa7dd8a9b4ac159b3"} Dec 04 17:53:37 crc kubenswrapper[4733]: I1204 17:53:37.067786 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ec5f4dab1b4de97e83e8ad8ac525bac9abb9f296786376fa7dd8a9b4ac159b3" Dec 04 17:53:37 crc kubenswrapper[4733]: I1204 17:53:37.067833 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.495610 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln"] Dec 04 17:53:45 crc kubenswrapper[4733]: E1204 17:53:45.496045 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerName="extract" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.496057 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerName="extract" Dec 04 17:53:45 crc kubenswrapper[4733]: E1204 17:53:45.496075 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerName="util" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.496080 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerName="util" Dec 04 17:53:45 crc kubenswrapper[4733]: E1204 17:53:45.496092 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerName="pull" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.496098 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerName="pull" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.496187 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d54cd1df-e504-4d88-8d0d-41610c92c294" containerName="extract" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.496513 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.503899 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.503901 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.503966 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-pf9nd" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.503969 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.504033 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.520997 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln"] Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.663346 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b8339d9-7351-422c-a9f3-4d9859dcff2f-webhook-cert\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.663462 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8lsr\" (UniqueName: \"kubernetes.io/projected/3b8339d9-7351-422c-a9f3-4d9859dcff2f-kube-api-access-t8lsr\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.663488 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b8339d9-7351-422c-a9f3-4d9859dcff2f-apiservice-cert\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.756027 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g"] Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.756766 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.759005 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8dfjx" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.759577 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.759608 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.764499 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8lsr\" (UniqueName: \"kubernetes.io/projected/3b8339d9-7351-422c-a9f3-4d9859dcff2f-kube-api-access-t8lsr\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.764547 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b8339d9-7351-422c-a9f3-4d9859dcff2f-apiservice-cert\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.764594 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b8339d9-7351-422c-a9f3-4d9859dcff2f-webhook-cert\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.764967 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g"] Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.780989 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b8339d9-7351-422c-a9f3-4d9859dcff2f-apiservice-cert\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.788454 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b8339d9-7351-422c-a9f3-4d9859dcff2f-webhook-cert\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.790659 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8lsr\" (UniqueName: \"kubernetes.io/projected/3b8339d9-7351-422c-a9f3-4d9859dcff2f-kube-api-access-t8lsr\") pod \"metallb-operator-controller-manager-77769ff6c4-bp6ln\" (UID: \"3b8339d9-7351-422c-a9f3-4d9859dcff2f\") " pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.810570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.865610 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4s8p\" (UniqueName: \"kubernetes.io/projected/e3cd2ad2-42a7-42b1-adb6-0d6080922788-kube-api-access-c4s8p\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.865710 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e3cd2ad2-42a7-42b1-adb6-0d6080922788-webhook-cert\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.865744 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e3cd2ad2-42a7-42b1-adb6-0d6080922788-apiservice-cert\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.967001 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4s8p\" (UniqueName: \"kubernetes.io/projected/e3cd2ad2-42a7-42b1-adb6-0d6080922788-kube-api-access-c4s8p\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.967075 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e3cd2ad2-42a7-42b1-adb6-0d6080922788-webhook-cert\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.967099 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e3cd2ad2-42a7-42b1-adb6-0d6080922788-apiservice-cert\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.974387 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e3cd2ad2-42a7-42b1-adb6-0d6080922788-webhook-cert\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:45 crc kubenswrapper[4733]: I1204 17:53:45.986122 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e3cd2ad2-42a7-42b1-adb6-0d6080922788-apiservice-cert\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:46 crc kubenswrapper[4733]: I1204 17:53:46.005854 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4s8p\" (UniqueName: \"kubernetes.io/projected/e3cd2ad2-42a7-42b1-adb6-0d6080922788-kube-api-access-c4s8p\") pod \"metallb-operator-webhook-server-7fc6b79947-ljf4g\" (UID: \"e3cd2ad2-42a7-42b1-adb6-0d6080922788\") " pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:46 crc kubenswrapper[4733]: I1204 17:53:46.071543 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:46 crc kubenswrapper[4733]: I1204 17:53:46.152424 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln"] Dec 04 17:53:46 crc kubenswrapper[4733]: W1204 17:53:46.159596 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b8339d9_7351_422c_a9f3_4d9859dcff2f.slice/crio-5cc7dc27df9e660d41486ce50b9f498a0d5e8486610676dfa59d480198479540 WatchSource:0}: Error finding container 5cc7dc27df9e660d41486ce50b9f498a0d5e8486610676dfa59d480198479540: Status 404 returned error can't find the container with id 5cc7dc27df9e660d41486ce50b9f498a0d5e8486610676dfa59d480198479540 Dec 04 17:53:46 crc kubenswrapper[4733]: I1204 17:53:46.494567 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g"] Dec 04 17:53:46 crc kubenswrapper[4733]: W1204 17:53:46.497693 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3cd2ad2_42a7_42b1_adb6_0d6080922788.slice/crio-d3fdc247fd92fe8feac433846b3f8e03ae54d034a4946c4bfbb09866daf8b5fa WatchSource:0}: Error finding container d3fdc247fd92fe8feac433846b3f8e03ae54d034a4946c4bfbb09866daf8b5fa: Status 404 returned error can't find the container with id d3fdc247fd92fe8feac433846b3f8e03ae54d034a4946c4bfbb09866daf8b5fa Dec 04 17:53:47 crc kubenswrapper[4733]: I1204 17:53:47.130925 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" event={"ID":"3b8339d9-7351-422c-a9f3-4d9859dcff2f","Type":"ContainerStarted","Data":"5cc7dc27df9e660d41486ce50b9f498a0d5e8486610676dfa59d480198479540"} Dec 04 17:53:47 crc kubenswrapper[4733]: I1204 17:53:47.132139 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" event={"ID":"e3cd2ad2-42a7-42b1-adb6-0d6080922788","Type":"ContainerStarted","Data":"d3fdc247fd92fe8feac433846b3f8e03ae54d034a4946c4bfbb09866daf8b5fa"} Dec 04 17:53:51 crc kubenswrapper[4733]: I1204 17:53:51.156306 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" event={"ID":"3b8339d9-7351-422c-a9f3-4d9859dcff2f","Type":"ContainerStarted","Data":"ceffe60f947e70c7a824be1f3f8e6a67bf9920c5639c8fa1808c6350fbe36182"} Dec 04 17:53:51 crc kubenswrapper[4733]: I1204 17:53:51.156747 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:53:51 crc kubenswrapper[4733]: I1204 17:53:51.158157 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" event={"ID":"e3cd2ad2-42a7-42b1-adb6-0d6080922788","Type":"ContainerStarted","Data":"18eacc0c93ce8004580021dd0ec2ca316e34f70a296e7121b5ae884ad3924ea8"} Dec 04 17:53:51 crc kubenswrapper[4733]: I1204 17:53:51.158476 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:53:51 crc kubenswrapper[4733]: I1204 17:53:51.181043 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" podStartSLOduration=2.086391515 podStartE2EDuration="6.181009979s" podCreationTimestamp="2025-12-04 17:53:45 +0000 UTC" firstStartedPulling="2025-12-04 17:53:46.162341337 +0000 UTC m=+888.117702383" lastFinishedPulling="2025-12-04 17:53:50.256959801 +0000 UTC m=+892.212320847" observedRunningTime="2025-12-04 17:53:51.178613604 +0000 UTC m=+893.133974660" watchObservedRunningTime="2025-12-04 17:53:51.181009979 +0000 UTC m=+893.136371065" Dec 04 17:53:51 crc kubenswrapper[4733]: I1204 17:53:51.208013 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" podStartSLOduration=2.413383609 podStartE2EDuration="6.20798945s" podCreationTimestamp="2025-12-04 17:53:45 +0000 UTC" firstStartedPulling="2025-12-04 17:53:46.500173965 +0000 UTC m=+888.455535011" lastFinishedPulling="2025-12-04 17:53:50.294779806 +0000 UTC m=+892.250140852" observedRunningTime="2025-12-04 17:53:51.205403281 +0000 UTC m=+893.160764337" watchObservedRunningTime="2025-12-04 17:53:51.20798945 +0000 UTC m=+893.163350526" Dec 04 17:54:06 crc kubenswrapper[4733]: I1204 17:54:06.097591 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7fc6b79947-ljf4g" Dec 04 17:54:25 crc kubenswrapper[4733]: I1204 17:54:25.814321 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-77769ff6c4-bp6ln" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.604458 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-mrbzw"] Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.607392 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.609132 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xb4bq" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.609460 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.610447 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.611413 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr"] Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.612076 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.614501 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.624346 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr"] Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.696888 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-lqqdn"] Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.697953 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.699717 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-pr2jg" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.699809 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.699885 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.700020 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.717025 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-z54dn"] Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.717923 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719582 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-conf\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719610 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-startup\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719635 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d64kp\" (UniqueName: \"kubernetes.io/projected/1ea286a4-29bd-4270-ba70-f22bed08fe14-kube-api-access-d64kp\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719658 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-metrics\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719675 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-reloader\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719693 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-sockets\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719708 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-nb7dr\" (UID: \"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719724 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ea286a4-29bd-4270-ba70-f22bed08fe14-metrics-certs\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719760 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbcn9\" (UniqueName: \"kubernetes.io/projected/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-kube-api-access-bbcn9\") pod \"frr-k8s-webhook-server-7fcb986d4-nb7dr\" (UID: \"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.719970 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.728997 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-z54dn"] Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.820988 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-startup\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821032 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-cert\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821062 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d64kp\" (UniqueName: \"kubernetes.io/projected/1ea286a4-29bd-4270-ba70-f22bed08fe14-kube-api-access-d64kp\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821085 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd72q\" (UniqueName: \"kubernetes.io/projected/d1b22abb-0421-41f8-a18e-89ddc9685791-kube-api-access-jd72q\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821106 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metrics-certs\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821124 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-metrics\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821144 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821162 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-reloader\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821178 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-sockets\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821244 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-nb7dr\" (UID: \"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821263 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ea286a4-29bd-4270-ba70-f22bed08fe14-metrics-certs\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821283 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metallb-excludel2\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821338 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-metrics-certs\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821366 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbcn9\" (UniqueName: \"kubernetes.io/projected/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-kube-api-access-bbcn9\") pod \"frr-k8s-webhook-server-7fcb986d4-nb7dr\" (UID: \"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821388 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jgvv\" (UniqueName: \"kubernetes.io/projected/c76f0cb9-9087-4c60-8152-4bdccc3862ef-kube-api-access-8jgvv\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821408 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-conf\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: E1204 17:54:26.821640 4733 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 04 17:54:26 crc kubenswrapper[4733]: E1204 17:54:26.821719 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-cert podName:f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f nodeName:}" failed. No retries permitted until 2025-12-04 17:54:27.321702688 +0000 UTC m=+929.277063844 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-cert") pod "frr-k8s-webhook-server-7fcb986d4-nb7dr" (UID: "f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f") : secret "frr-k8s-webhook-server-cert" not found Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.821832 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-metrics\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.822046 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-conf\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.822095 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-sockets\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.822125 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ea286a4-29bd-4270-ba70-f22bed08fe14-reloader\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.822271 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ea286a4-29bd-4270-ba70-f22bed08fe14-frr-startup\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.830350 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ea286a4-29bd-4270-ba70-f22bed08fe14-metrics-certs\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.839347 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d64kp\" (UniqueName: \"kubernetes.io/projected/1ea286a4-29bd-4270-ba70-f22bed08fe14-kube-api-access-d64kp\") pod \"frr-k8s-mrbzw\" (UID: \"1ea286a4-29bd-4270-ba70-f22bed08fe14\") " pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.842057 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbcn9\" (UniqueName: \"kubernetes.io/projected/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-kube-api-access-bbcn9\") pod \"frr-k8s-webhook-server-7fcb986d4-nb7dr\" (UID: \"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.923183 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metallb-excludel2\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.923280 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-metrics-certs\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.923332 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jgvv\" (UniqueName: \"kubernetes.io/projected/c76f0cb9-9087-4c60-8152-4bdccc3862ef-kube-api-access-8jgvv\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.923385 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-cert\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.923440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd72q\" (UniqueName: \"kubernetes.io/projected/d1b22abb-0421-41f8-a18e-89ddc9685791-kube-api-access-jd72q\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.923478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metrics-certs\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.923513 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: E1204 17:54:26.923688 4733 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 17:54:26 crc kubenswrapper[4733]: E1204 17:54:26.923762 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist podName:c76f0cb9-9087-4c60-8152-4bdccc3862ef nodeName:}" failed. No retries permitted until 2025-12-04 17:54:27.423739485 +0000 UTC m=+929.379100551 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist") pod "speaker-lqqdn" (UID: "c76f0cb9-9087-4c60-8152-4bdccc3862ef") : secret "metallb-memberlist" not found Dec 04 17:54:26 crc kubenswrapper[4733]: E1204 17:54:26.925054 4733 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.925098 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metallb-excludel2\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: E1204 17:54:26.925135 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-metrics-certs podName:d1b22abb-0421-41f8-a18e-89ddc9685791 nodeName:}" failed. No retries permitted until 2025-12-04 17:54:27.425116612 +0000 UTC m=+929.380477658 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-metrics-certs") pod "controller-f8648f98b-z54dn" (UID: "d1b22abb-0421-41f8-a18e-89ddc9685791") : secret "controller-certs-secret" not found Dec 04 17:54:26 crc kubenswrapper[4733]: E1204 17:54:26.925454 4733 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 04 17:54:26 crc kubenswrapper[4733]: E1204 17:54:26.925491 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metrics-certs podName:c76f0cb9-9087-4c60-8152-4bdccc3862ef nodeName:}" failed. No retries permitted until 2025-12-04 17:54:27.425481982 +0000 UTC m=+929.380843028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metrics-certs") pod "speaker-lqqdn" (UID: "c76f0cb9-9087-4c60-8152-4bdccc3862ef") : secret "speaker-certs-secret" not found Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.926624 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.930651 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.939300 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-cert\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.942479 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jgvv\" (UniqueName: \"kubernetes.io/projected/c76f0cb9-9087-4c60-8152-4bdccc3862ef-kube-api-access-8jgvv\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:26 crc kubenswrapper[4733]: I1204 17:54:26.951834 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd72q\" (UniqueName: \"kubernetes.io/projected/d1b22abb-0421-41f8-a18e-89ddc9685791-kube-api-access-jd72q\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.329928 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-nb7dr\" (UID: \"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.336608 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-nb7dr\" (UID: \"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.402539 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerStarted","Data":"5a9bc2a5c78144fd17253e5a144ceec9164f87ba4832f3343eb101cb826f7cde"} Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.431759 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-metrics-certs\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.431942 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metrics-certs\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.431982 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:27 crc kubenswrapper[4733]: E1204 17:54:27.432156 4733 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 17:54:27 crc kubenswrapper[4733]: E1204 17:54:27.432822 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist podName:c76f0cb9-9087-4c60-8152-4bdccc3862ef nodeName:}" failed. No retries permitted until 2025-12-04 17:54:28.432762144 +0000 UTC m=+930.388123220 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist") pod "speaker-lqqdn" (UID: "c76f0cb9-9087-4c60-8152-4bdccc3862ef") : secret "metallb-memberlist" not found Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.436507 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d1b22abb-0421-41f8-a18e-89ddc9685791-metrics-certs\") pod \"controller-f8648f98b-z54dn\" (UID: \"d1b22abb-0421-41f8-a18e-89ddc9685791\") " pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.437571 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-metrics-certs\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.541130 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.631399 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.832403 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr"] Dec 04 17:54:27 crc kubenswrapper[4733]: I1204 17:54:27.916494 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-z54dn"] Dec 04 17:54:27 crc kubenswrapper[4733]: W1204 17:54:27.923370 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b22abb_0421_41f8_a18e_89ddc9685791.slice/crio-01160ce839e4cde7dd9b32e894711872b8a12accd1eb41737d198f87c8017916 WatchSource:0}: Error finding container 01160ce839e4cde7dd9b32e894711872b8a12accd1eb41737d198f87c8017916: Status 404 returned error can't find the container with id 01160ce839e4cde7dd9b32e894711872b8a12accd1eb41737d198f87c8017916 Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.421283 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" event={"ID":"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f","Type":"ContainerStarted","Data":"abb198a8443e3efe2861c068b445d1b527c98062e13f2f8c82821ea2d4f29baa"} Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.423494 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-z54dn" event={"ID":"d1b22abb-0421-41f8-a18e-89ddc9685791","Type":"ContainerStarted","Data":"f3f94137ab8def42bd7b7a6fc50c14f451f9550379070b19c2abb12253cbfae4"} Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.423520 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-z54dn" event={"ID":"d1b22abb-0421-41f8-a18e-89ddc9685791","Type":"ContainerStarted","Data":"4578a5d8f4729187e243aa08fc68255c9367bef39586452f54e057986b7dab96"} Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.423530 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-z54dn" event={"ID":"d1b22abb-0421-41f8-a18e-89ddc9685791","Type":"ContainerStarted","Data":"01160ce839e4cde7dd9b32e894711872b8a12accd1eb41737d198f87c8017916"} Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.423676 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.445472 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-z54dn" podStartSLOduration=2.445456555 podStartE2EDuration="2.445456555s" podCreationTimestamp="2025-12-04 17:54:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:54:28.439419741 +0000 UTC m=+930.394780797" watchObservedRunningTime="2025-12-04 17:54:28.445456555 +0000 UTC m=+930.400817601" Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.450360 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.455400 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c76f0cb9-9087-4c60-8152-4bdccc3862ef-memberlist\") pod \"speaker-lqqdn\" (UID: \"c76f0cb9-9087-4c60-8152-4bdccc3862ef\") " pod="metallb-system/speaker-lqqdn" Dec 04 17:54:28 crc kubenswrapper[4733]: I1204 17:54:28.511105 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-lqqdn" Dec 04 17:54:28 crc kubenswrapper[4733]: W1204 17:54:28.529499 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc76f0cb9_9087_4c60_8152_4bdccc3862ef.slice/crio-2a916b52d30f799c5ea45c84861d308c54bdf3ada22bc3ee28552a77ab6c0792 WatchSource:0}: Error finding container 2a916b52d30f799c5ea45c84861d308c54bdf3ada22bc3ee28552a77ab6c0792: Status 404 returned error can't find the container with id 2a916b52d30f799c5ea45c84861d308c54bdf3ada22bc3ee28552a77ab6c0792 Dec 04 17:54:29 crc kubenswrapper[4733]: I1204 17:54:29.437446 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lqqdn" event={"ID":"c76f0cb9-9087-4c60-8152-4bdccc3862ef","Type":"ContainerStarted","Data":"b2871fe003893dbee0e2595e9ddb1c29a72d8c2b3d0c941f0b3cb674899184be"} Dec 04 17:54:29 crc kubenswrapper[4733]: I1204 17:54:29.437766 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lqqdn" event={"ID":"c76f0cb9-9087-4c60-8152-4bdccc3862ef","Type":"ContainerStarted","Data":"e05a14ac4a068194f64cc86e3709e36a5cc1416960b69039de859b13406114b1"} Dec 04 17:54:29 crc kubenswrapper[4733]: I1204 17:54:29.437781 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lqqdn" event={"ID":"c76f0cb9-9087-4c60-8152-4bdccc3862ef","Type":"ContainerStarted","Data":"2a916b52d30f799c5ea45c84861d308c54bdf3ada22bc3ee28552a77ab6c0792"} Dec 04 17:54:29 crc kubenswrapper[4733]: I1204 17:54:29.438124 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-lqqdn" Dec 04 17:54:29 crc kubenswrapper[4733]: I1204 17:54:29.457504 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-lqqdn" podStartSLOduration=3.457485078 podStartE2EDuration="3.457485078s" podCreationTimestamp="2025-12-04 17:54:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:54:29.45350418 +0000 UTC m=+931.408865226" watchObservedRunningTime="2025-12-04 17:54:29.457485078 +0000 UTC m=+931.412846114" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.327913 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mpt7c"] Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.329540 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.334135 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpt7c"] Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.403759 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-catalog-content\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.403838 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-utilities\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.403867 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4656g\" (UniqueName: \"kubernetes.io/projected/c83a12a3-b371-4845-81f2-90cbcef4793a-kube-api-access-4656g\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.505663 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-catalog-content\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.505717 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-utilities\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.505739 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4656g\" (UniqueName: \"kubernetes.io/projected/c83a12a3-b371-4845-81f2-90cbcef4793a-kube-api-access-4656g\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.506252 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-catalog-content\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.506430 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-utilities\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.526957 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4656g\" (UniqueName: \"kubernetes.io/projected/c83a12a3-b371-4845-81f2-90cbcef4793a-kube-api-access-4656g\") pod \"redhat-marketplace-mpt7c\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:31 crc kubenswrapper[4733]: I1204 17:54:31.655388 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:34 crc kubenswrapper[4733]: I1204 17:54:34.307783 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpt7c"] Dec 04 17:54:34 crc kubenswrapper[4733]: I1204 17:54:34.475099 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpt7c" event={"ID":"c83a12a3-b371-4845-81f2-90cbcef4793a","Type":"ContainerStarted","Data":"6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28"} Dec 04 17:54:34 crc kubenswrapper[4733]: I1204 17:54:34.475153 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpt7c" event={"ID":"c83a12a3-b371-4845-81f2-90cbcef4793a","Type":"ContainerStarted","Data":"b7bd1ad404ad627dab24fd3e99e8f4de62c390dd920721dbc9d6849d86cb008d"} Dec 04 17:54:34 crc kubenswrapper[4733]: I1204 17:54:34.476550 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" event={"ID":"f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f","Type":"ContainerStarted","Data":"948576e3fa4f8268a9ae7ecf8406fd6fa416e8f8541b33920bc416c3afcb330d"} Dec 04 17:54:34 crc kubenswrapper[4733]: I1204 17:54:34.476720 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:34 crc kubenswrapper[4733]: I1204 17:54:34.482146 4733 generic.go:334] "Generic (PLEG): container finished" podID="1ea286a4-29bd-4270-ba70-f22bed08fe14" containerID="05f0274f80b654542400c6df3d316d21254bf1d1f678272c67ef230d4509a8db" exitCode=0 Dec 04 17:54:34 crc kubenswrapper[4733]: I1204 17:54:34.482301 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerDied","Data":"05f0274f80b654542400c6df3d316d21254bf1d1f678272c67ef230d4509a8db"} Dec 04 17:54:35 crc kubenswrapper[4733]: I1204 17:54:35.492207 4733 generic.go:334] "Generic (PLEG): container finished" podID="1ea286a4-29bd-4270-ba70-f22bed08fe14" containerID="12261be2e0ca7e7644c09b095dc72ebc957750c5f29ffd27d150cfc1c0aecb6f" exitCode=0 Dec 04 17:54:35 crc kubenswrapper[4733]: I1204 17:54:35.492340 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerDied","Data":"12261be2e0ca7e7644c09b095dc72ebc957750c5f29ffd27d150cfc1c0aecb6f"} Dec 04 17:54:35 crc kubenswrapper[4733]: I1204 17:54:35.494238 4733 generic.go:334] "Generic (PLEG): container finished" podID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerID="6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28" exitCode=0 Dec 04 17:54:35 crc kubenswrapper[4733]: I1204 17:54:35.494313 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpt7c" event={"ID":"c83a12a3-b371-4845-81f2-90cbcef4793a","Type":"ContainerDied","Data":"6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28"} Dec 04 17:54:35 crc kubenswrapper[4733]: I1204 17:54:35.543728 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" podStartSLOduration=3.457811567 podStartE2EDuration="9.543704899s" podCreationTimestamp="2025-12-04 17:54:26 +0000 UTC" firstStartedPulling="2025-12-04 17:54:27.843525038 +0000 UTC m=+929.798886094" lastFinishedPulling="2025-12-04 17:54:33.92941837 +0000 UTC m=+935.884779426" observedRunningTime="2025-12-04 17:54:34.547956546 +0000 UTC m=+936.503317642" watchObservedRunningTime="2025-12-04 17:54:35.543704899 +0000 UTC m=+937.499065985" Dec 04 17:54:36 crc kubenswrapper[4733]: I1204 17:54:36.501905 4733 generic.go:334] "Generic (PLEG): container finished" podID="1ea286a4-29bd-4270-ba70-f22bed08fe14" containerID="458d2b03b0c79591005682f4bbe271cf08d0c9632f8bf2135aeb2949c68bc14c" exitCode=0 Dec 04 17:54:36 crc kubenswrapper[4733]: I1204 17:54:36.501971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerDied","Data":"458d2b03b0c79591005682f4bbe271cf08d0c9632f8bf2135aeb2949c68bc14c"} Dec 04 17:54:36 crc kubenswrapper[4733]: I1204 17:54:36.504587 4733 generic.go:334] "Generic (PLEG): container finished" podID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerID="a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2" exitCode=0 Dec 04 17:54:36 crc kubenswrapper[4733]: I1204 17:54:36.504654 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpt7c" event={"ID":"c83a12a3-b371-4845-81f2-90cbcef4793a","Type":"ContainerDied","Data":"a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2"} Dec 04 17:54:37 crc kubenswrapper[4733]: I1204 17:54:37.514524 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerStarted","Data":"f8f1aef6f8eba8811b230c16a03b15ae5aca0f7c2d5b30ff7b64344f3183852f"} Dec 04 17:54:37 crc kubenswrapper[4733]: I1204 17:54:37.514579 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerStarted","Data":"bf852f3bea8525a4e2f7ed0f1a6cf6a3fed413ca486dbc7f25d4df3f2e3e1454"} Dec 04 17:54:37 crc kubenswrapper[4733]: I1204 17:54:37.514593 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerStarted","Data":"1e0c005527eddcf9e103ab56918ad7d678eef8c53d9b1a5dad433fb44ec0ac15"} Dec 04 17:54:37 crc kubenswrapper[4733]: I1204 17:54:37.514610 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerStarted","Data":"3cef951222306bda559985eae23ee18c566d2a0c9a4187a688a940757d76d79b"} Dec 04 17:54:37 crc kubenswrapper[4733]: I1204 17:54:37.514624 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerStarted","Data":"3bf4e9699edc6e75e6e5aeae0135535d729b6cf4a6b70bf4ec24a9645e4e4bdf"} Dec 04 17:54:37 crc kubenswrapper[4733]: I1204 17:54:37.516642 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpt7c" event={"ID":"c83a12a3-b371-4845-81f2-90cbcef4793a","Type":"ContainerStarted","Data":"f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac"} Dec 04 17:54:37 crc kubenswrapper[4733]: I1204 17:54:37.533136 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mpt7c" podStartSLOduration=5.044986047 podStartE2EDuration="6.533118276s" podCreationTimestamp="2025-12-04 17:54:31 +0000 UTC" firstStartedPulling="2025-12-04 17:54:35.496845568 +0000 UTC m=+937.452206654" lastFinishedPulling="2025-12-04 17:54:36.984977797 +0000 UTC m=+938.940338883" observedRunningTime="2025-12-04 17:54:37.531480491 +0000 UTC m=+939.486841547" watchObservedRunningTime="2025-12-04 17:54:37.533118276 +0000 UTC m=+939.488479332" Dec 04 17:54:38 crc kubenswrapper[4733]: I1204 17:54:38.515409 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-lqqdn" Dec 04 17:54:38 crc kubenswrapper[4733]: I1204 17:54:38.529720 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mrbzw" event={"ID":"1ea286a4-29bd-4270-ba70-f22bed08fe14","Type":"ContainerStarted","Data":"339fc282a608fbe6a05adc83ff39571748a49be8cd46bac76807e657c920d9b6"} Dec 04 17:54:38 crc kubenswrapper[4733]: I1204 17:54:38.585441 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-mrbzw" podStartSLOduration=5.773528549 podStartE2EDuration="12.585424541s" podCreationTimestamp="2025-12-04 17:54:26 +0000 UTC" firstStartedPulling="2025-12-04 17:54:27.095510161 +0000 UTC m=+929.050871247" lastFinishedPulling="2025-12-04 17:54:33.907406173 +0000 UTC m=+935.862767239" observedRunningTime="2025-12-04 17:54:38.577491386 +0000 UTC m=+940.532852442" watchObservedRunningTime="2025-12-04 17:54:38.585424541 +0000 UTC m=+940.540785587" Dec 04 17:54:39 crc kubenswrapper[4733]: I1204 17:54:39.536848 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.015933 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj"] Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.018061 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.019950 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.028410 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj"] Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.115175 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62s28\" (UniqueName: \"kubernetes.io/projected/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-kube-api-access-62s28\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.115255 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.115296 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.216470 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62s28\" (UniqueName: \"kubernetes.io/projected/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-kube-api-access-62s28\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.216581 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.216620 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.217316 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.217354 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.233962 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62s28\" (UniqueName: \"kubernetes.io/projected/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-kube-api-access-62s28\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.333275 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:40 crc kubenswrapper[4733]: I1204 17:54:40.812430 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj"] Dec 04 17:54:40 crc kubenswrapper[4733]: W1204 17:54:40.819336 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aaba4f3_8c8c_41e8_b7ca_1fc389f1f512.slice/crio-4bc285d61fd7918ea7417dd5af578107f1b40ea6319a1d422e687d7e1374ee8c WatchSource:0}: Error finding container 4bc285d61fd7918ea7417dd5af578107f1b40ea6319a1d422e687d7e1374ee8c: Status 404 returned error can't find the container with id 4bc285d61fd7918ea7417dd5af578107f1b40ea6319a1d422e687d7e1374ee8c Dec 04 17:54:41 crc kubenswrapper[4733]: I1204 17:54:41.548619 4733 generic.go:334] "Generic (PLEG): container finished" podID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerID="d3f3514522c70e60e0514299027f20ddfe7f34219e989026c604baf84d7c9869" exitCode=0 Dec 04 17:54:41 crc kubenswrapper[4733]: I1204 17:54:41.548716 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" event={"ID":"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512","Type":"ContainerDied","Data":"d3f3514522c70e60e0514299027f20ddfe7f34219e989026c604baf84d7c9869"} Dec 04 17:54:41 crc kubenswrapper[4733]: I1204 17:54:41.548953 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" event={"ID":"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512","Type":"ContainerStarted","Data":"4bc285d61fd7918ea7417dd5af578107f1b40ea6319a1d422e687d7e1374ee8c"} Dec 04 17:54:41 crc kubenswrapper[4733]: I1204 17:54:41.656114 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:41 crc kubenswrapper[4733]: I1204 17:54:41.656149 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:41 crc kubenswrapper[4733]: I1204 17:54:41.711590 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:41 crc kubenswrapper[4733]: I1204 17:54:41.932076 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:41 crc kubenswrapper[4733]: I1204 17:54:41.970478 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:42 crc kubenswrapper[4733]: I1204 17:54:42.607070 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:44 crc kubenswrapper[4733]: I1204 17:54:44.163759 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpt7c"] Dec 04 17:54:44 crc kubenswrapper[4733]: I1204 17:54:44.573738 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mpt7c" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerName="registry-server" containerID="cri-o://f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac" gracePeriod=2 Dec 04 17:54:44 crc kubenswrapper[4733]: I1204 17:54:44.951494 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.106713 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4656g\" (UniqueName: \"kubernetes.io/projected/c83a12a3-b371-4845-81f2-90cbcef4793a-kube-api-access-4656g\") pod \"c83a12a3-b371-4845-81f2-90cbcef4793a\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.106895 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-catalog-content\") pod \"c83a12a3-b371-4845-81f2-90cbcef4793a\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.107007 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-utilities\") pod \"c83a12a3-b371-4845-81f2-90cbcef4793a\" (UID: \"c83a12a3-b371-4845-81f2-90cbcef4793a\") " Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.107746 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-utilities" (OuterVolumeSpecName: "utilities") pod "c83a12a3-b371-4845-81f2-90cbcef4793a" (UID: "c83a12a3-b371-4845-81f2-90cbcef4793a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.117477 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c83a12a3-b371-4845-81f2-90cbcef4793a-kube-api-access-4656g" (OuterVolumeSpecName: "kube-api-access-4656g") pod "c83a12a3-b371-4845-81f2-90cbcef4793a" (UID: "c83a12a3-b371-4845-81f2-90cbcef4793a"). InnerVolumeSpecName "kube-api-access-4656g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.133186 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c83a12a3-b371-4845-81f2-90cbcef4793a" (UID: "c83a12a3-b371-4845-81f2-90cbcef4793a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.208609 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4656g\" (UniqueName: \"kubernetes.io/projected/c83a12a3-b371-4845-81f2-90cbcef4793a-kube-api-access-4656g\") on node \"crc\" DevicePath \"\"" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.208653 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.208665 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83a12a3-b371-4845-81f2-90cbcef4793a-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.588006 4733 generic.go:334] "Generic (PLEG): container finished" podID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerID="f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac" exitCode=0 Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.588084 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpt7c" event={"ID":"c83a12a3-b371-4845-81f2-90cbcef4793a","Type":"ContainerDied","Data":"f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac"} Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.588107 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpt7c" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.588575 4733 scope.go:117] "RemoveContainer" containerID="f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.588546 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpt7c" event={"ID":"c83a12a3-b371-4845-81f2-90cbcef4793a","Type":"ContainerDied","Data":"b7bd1ad404ad627dab24fd3e99e8f4de62c390dd920721dbc9d6849d86cb008d"} Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.594576 4733 generic.go:334] "Generic (PLEG): container finished" podID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerID="73038c127468225ba302a310420d4b93a38630d9cd425f3c13b091fff887499d" exitCode=0 Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.594623 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" event={"ID":"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512","Type":"ContainerDied","Data":"73038c127468225ba302a310420d4b93a38630d9cd425f3c13b091fff887499d"} Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.640184 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpt7c"] Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.641063 4733 scope.go:117] "RemoveContainer" containerID="a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.646478 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpt7c"] Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.690990 4733 scope.go:117] "RemoveContainer" containerID="6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.737907 4733 scope.go:117] "RemoveContainer" containerID="f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac" Dec 04 17:54:45 crc kubenswrapper[4733]: E1204 17:54:45.738783 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac\": container with ID starting with f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac not found: ID does not exist" containerID="f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.738887 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac"} err="failed to get container status \"f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac\": rpc error: code = NotFound desc = could not find container \"f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac\": container with ID starting with f76f3406be9ef36463000577122c643b69e6853d81b40e43545a04a592336dac not found: ID does not exist" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.738920 4733 scope.go:117] "RemoveContainer" containerID="a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2" Dec 04 17:54:45 crc kubenswrapper[4733]: E1204 17:54:45.739485 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2\": container with ID starting with a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2 not found: ID does not exist" containerID="a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.739526 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2"} err="failed to get container status \"a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2\": rpc error: code = NotFound desc = could not find container \"a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2\": container with ID starting with a50e49b703cf4f5808a171063696162e9935426bbaf603b1012c67db5d2079f2 not found: ID does not exist" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.739552 4733 scope.go:117] "RemoveContainer" containerID="6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28" Dec 04 17:54:45 crc kubenswrapper[4733]: E1204 17:54:45.739985 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28\": container with ID starting with 6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28 not found: ID does not exist" containerID="6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28" Dec 04 17:54:45 crc kubenswrapper[4733]: I1204 17:54:45.740018 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28"} err="failed to get container status \"6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28\": rpc error: code = NotFound desc = could not find container \"6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28\": container with ID starting with 6624318e0bbfbbcc6663433785ab73b95b53102a5b39e180483be305465fbc28 not found: ID does not exist" Dec 04 17:54:46 crc kubenswrapper[4733]: I1204 17:54:46.345986 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" path="/var/lib/kubelet/pods/c83a12a3-b371-4845-81f2-90cbcef4793a/volumes" Dec 04 17:54:46 crc kubenswrapper[4733]: I1204 17:54:46.609216 4733 generic.go:334] "Generic (PLEG): container finished" podID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerID="79e99ca0c8f9df8c35755ea4cf8213f619fa69a38d3f50d76fac161f30ae48ef" exitCode=0 Dec 04 17:54:46 crc kubenswrapper[4733]: I1204 17:54:46.609260 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" event={"ID":"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512","Type":"ContainerDied","Data":"79e99ca0c8f9df8c35755ea4cf8213f619fa69a38d3f50d76fac161f30ae48ef"} Dec 04 17:54:46 crc kubenswrapper[4733]: I1204 17:54:46.934898 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-mrbzw" Dec 04 17:54:47 crc kubenswrapper[4733]: I1204 17:54:47.546906 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-nb7dr" Dec 04 17:54:47 crc kubenswrapper[4733]: I1204 17:54:47.637333 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-z54dn" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.158912 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.273223 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-bundle\") pod \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.273349 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-util\") pod \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.273379 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62s28\" (UniqueName: \"kubernetes.io/projected/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-kube-api-access-62s28\") pod \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\" (UID: \"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512\") " Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.274692 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-bundle" (OuterVolumeSpecName: "bundle") pod "4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" (UID: "4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.279648 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-kube-api-access-62s28" (OuterVolumeSpecName: "kube-api-access-62s28") pod "4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" (UID: "4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512"). InnerVolumeSpecName "kube-api-access-62s28". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.284960 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-util" (OuterVolumeSpecName: "util") pod "4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" (UID: "4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.375523 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.375554 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-util\") on node \"crc\" DevicePath \"\"" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.375564 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62s28\" (UniqueName: \"kubernetes.io/projected/4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512-kube-api-access-62s28\") on node \"crc\" DevicePath \"\"" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.627434 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" event={"ID":"4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512","Type":"ContainerDied","Data":"4bc285d61fd7918ea7417dd5af578107f1b40ea6319a1d422e687d7e1374ee8c"} Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.627507 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bc285d61fd7918ea7417dd5af578107f1b40ea6319a1d422e687d7e1374ee8c" Dec 04 17:54:48 crc kubenswrapper[4733]: I1204 17:54:48.627509 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.602469 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7"] Dec 04 17:54:52 crc kubenswrapper[4733]: E1204 17:54:52.603389 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerName="extract" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.603415 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerName="extract" Dec 04 17:54:52 crc kubenswrapper[4733]: E1204 17:54:52.603433 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerName="extract-utilities" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.603445 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerName="extract-utilities" Dec 04 17:54:52 crc kubenswrapper[4733]: E1204 17:54:52.603481 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerName="util" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.603494 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerName="util" Dec 04 17:54:52 crc kubenswrapper[4733]: E1204 17:54:52.603518 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerName="registry-server" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.603531 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerName="registry-server" Dec 04 17:54:52 crc kubenswrapper[4733]: E1204 17:54:52.603553 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerName="pull" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.603566 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerName="pull" Dec 04 17:54:52 crc kubenswrapper[4733]: E1204 17:54:52.603593 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerName="extract-content" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.603609 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerName="extract-content" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.603838 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512" containerName="extract" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.603868 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c83a12a3-b371-4845-81f2-90cbcef4793a" containerName="registry-server" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.604509 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.608276 4733 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-g2t4w" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.608629 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.608947 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.624270 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7"] Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.729884 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfcsw\" (UniqueName: \"kubernetes.io/projected/529939a5-aace-4c1e-b6ec-ff7685296f63-kube-api-access-tfcsw\") pod \"cert-manager-operator-controller-manager-64cf6dff88-gslc7\" (UID: \"529939a5-aace-4c1e-b6ec-ff7685296f63\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.729936 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/529939a5-aace-4c1e-b6ec-ff7685296f63-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-gslc7\" (UID: \"529939a5-aace-4c1e-b6ec-ff7685296f63\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.831760 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfcsw\" (UniqueName: \"kubernetes.io/projected/529939a5-aace-4c1e-b6ec-ff7685296f63-kube-api-access-tfcsw\") pod \"cert-manager-operator-controller-manager-64cf6dff88-gslc7\" (UID: \"529939a5-aace-4c1e-b6ec-ff7685296f63\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.831825 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/529939a5-aace-4c1e-b6ec-ff7685296f63-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-gslc7\" (UID: \"529939a5-aace-4c1e-b6ec-ff7685296f63\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.832307 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/529939a5-aace-4c1e-b6ec-ff7685296f63-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-gslc7\" (UID: \"529939a5-aace-4c1e-b6ec-ff7685296f63\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.867652 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfcsw\" (UniqueName: \"kubernetes.io/projected/529939a5-aace-4c1e-b6ec-ff7685296f63-kube-api-access-tfcsw\") pod \"cert-manager-operator-controller-manager-64cf6dff88-gslc7\" (UID: \"529939a5-aace-4c1e-b6ec-ff7685296f63\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" Dec 04 17:54:52 crc kubenswrapper[4733]: I1204 17:54:52.922578 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" Dec 04 17:54:53 crc kubenswrapper[4733]: I1204 17:54:53.196585 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7"] Dec 04 17:54:53 crc kubenswrapper[4733]: I1204 17:54:53.660670 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" event={"ID":"529939a5-aace-4c1e-b6ec-ff7685296f63","Type":"ContainerStarted","Data":"1bfe6efc3cb3a1bbb7bf00197da827da4d96cc7c75971b22a6c78a4ebafb01f7"} Dec 04 17:55:00 crc kubenswrapper[4733]: I1204 17:55:00.713920 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" event={"ID":"529939a5-aace-4c1e-b6ec-ff7685296f63","Type":"ContainerStarted","Data":"1b31a9b78f711ef637486c282af1e19dcbff34a796732543f4b8df7cb329db7d"} Dec 04 17:55:00 crc kubenswrapper[4733]: I1204 17:55:00.750141 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-gslc7" podStartSLOduration=1.890897429 podStartE2EDuration="8.750110714s" podCreationTimestamp="2025-12-04 17:54:52 +0000 UTC" firstStartedPulling="2025-12-04 17:54:53.20774917 +0000 UTC m=+955.163110216" lastFinishedPulling="2025-12-04 17:55:00.066962455 +0000 UTC m=+962.022323501" observedRunningTime="2025-12-04 17:55:00.745446347 +0000 UTC m=+962.700807453" watchObservedRunningTime="2025-12-04 17:55:00.750110714 +0000 UTC m=+962.705471830" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.244259 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-k5b27"] Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.245225 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.247477 4733 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-4v8sk" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.247535 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.247934 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.265523 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-k5b27"] Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.435436 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f22d997-ee5a-462e-8452-d493e4584761-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-k5b27\" (UID: \"8f22d997-ee5a-462e-8452-d493e4584761\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.435635 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb7q8\" (UniqueName: \"kubernetes.io/projected/8f22d997-ee5a-462e-8452-d493e4584761-kube-api-access-mb7q8\") pod \"cert-manager-webhook-f4fb5df64-k5b27\" (UID: \"8f22d997-ee5a-462e-8452-d493e4584761\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.536877 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb7q8\" (UniqueName: \"kubernetes.io/projected/8f22d997-ee5a-462e-8452-d493e4584761-kube-api-access-mb7q8\") pod \"cert-manager-webhook-f4fb5df64-k5b27\" (UID: \"8f22d997-ee5a-462e-8452-d493e4584761\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.537083 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f22d997-ee5a-462e-8452-d493e4584761-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-k5b27\" (UID: \"8f22d997-ee5a-462e-8452-d493e4584761\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.556525 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f22d997-ee5a-462e-8452-d493e4584761-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-k5b27\" (UID: \"8f22d997-ee5a-462e-8452-d493e4584761\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.558720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb7q8\" (UniqueName: \"kubernetes.io/projected/8f22d997-ee5a-462e-8452-d493e4584761-kube-api-access-mb7q8\") pod \"cert-manager-webhook-f4fb5df64-k5b27\" (UID: \"8f22d997-ee5a-462e-8452-d493e4584761\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.566424 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:05 crc kubenswrapper[4733]: I1204 17:55:05.790610 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-k5b27"] Dec 04 17:55:06 crc kubenswrapper[4733]: I1204 17:55:06.751348 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" event={"ID":"8f22d997-ee5a-462e-8452-d493e4584761","Type":"ContainerStarted","Data":"06596dc027b99af557c661117f50adcc5b8b73af3db9a568ee1f0969ebbeb374"} Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.741747 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm"] Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.742713 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.745965 4733 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-xlvzs" Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.748997 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm"] Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.869666 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtprp\" (UniqueName: \"kubernetes.io/projected/f2b4ae75-af8e-4f85-9139-2f4841ca4f94-kube-api-access-jtprp\") pod \"cert-manager-cainjector-855d9ccff4-wcdpm\" (UID: \"f2b4ae75-af8e-4f85-9139-2f4841ca4f94\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.869718 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2b4ae75-af8e-4f85-9139-2f4841ca4f94-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-wcdpm\" (UID: \"f2b4ae75-af8e-4f85-9139-2f4841ca4f94\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.971170 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtprp\" (UniqueName: \"kubernetes.io/projected/f2b4ae75-af8e-4f85-9139-2f4841ca4f94-kube-api-access-jtprp\") pod \"cert-manager-cainjector-855d9ccff4-wcdpm\" (UID: \"f2b4ae75-af8e-4f85-9139-2f4841ca4f94\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.971212 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2b4ae75-af8e-4f85-9139-2f4841ca4f94-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-wcdpm\" (UID: \"f2b4ae75-af8e-4f85-9139-2f4841ca4f94\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.991599 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtprp\" (UniqueName: \"kubernetes.io/projected/f2b4ae75-af8e-4f85-9139-2f4841ca4f94-kube-api-access-jtprp\") pod \"cert-manager-cainjector-855d9ccff4-wcdpm\" (UID: \"f2b4ae75-af8e-4f85-9139-2f4841ca4f94\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" Dec 04 17:55:07 crc kubenswrapper[4733]: I1204 17:55:07.995051 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2b4ae75-af8e-4f85-9139-2f4841ca4f94-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-wcdpm\" (UID: \"f2b4ae75-af8e-4f85-9139-2f4841ca4f94\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" Dec 04 17:55:08 crc kubenswrapper[4733]: I1204 17:55:08.066706 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" Dec 04 17:55:08 crc kubenswrapper[4733]: I1204 17:55:08.304262 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm"] Dec 04 17:55:08 crc kubenswrapper[4733]: I1204 17:55:08.772395 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" event={"ID":"f2b4ae75-af8e-4f85-9139-2f4841ca4f94","Type":"ContainerStarted","Data":"21514c9516b08a7696641d2e7f2cbbe6f522ac734e1eb062030acce4ac09c039"} Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.814846 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" event={"ID":"8f22d997-ee5a-462e-8452-d493e4584761","Type":"ContainerStarted","Data":"6f5004b3a0f262bd9e077b88d55391a16c2e081bc28d553b20de76f2500d286f"} Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.815486 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.817639 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" event={"ID":"f2b4ae75-af8e-4f85-9139-2f4841ca4f94","Type":"ContainerStarted","Data":"675ef35c2ee59bbb201ee6ee3de65bbf9af1d345674e2dfcb95b1eaf9c6d3efc"} Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.836543 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" podStartSLOduration=1.193255215 podStartE2EDuration="8.836520463s" podCreationTimestamp="2025-12-04 17:55:05 +0000 UTC" firstStartedPulling="2025-12-04 17:55:05.808005819 +0000 UTC m=+967.763366885" lastFinishedPulling="2025-12-04 17:55:13.451271037 +0000 UTC m=+975.406632133" observedRunningTime="2025-12-04 17:55:13.83084211 +0000 UTC m=+975.786203186" watchObservedRunningTime="2025-12-04 17:55:13.836520463 +0000 UTC m=+975.791881529" Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.853432 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-wcdpm" podStartSLOduration=1.834057286 podStartE2EDuration="6.853412807s" podCreationTimestamp="2025-12-04 17:55:07 +0000 UTC" firstStartedPulling="2025-12-04 17:55:08.345905204 +0000 UTC m=+970.301266250" lastFinishedPulling="2025-12-04 17:55:13.365260725 +0000 UTC m=+975.320621771" observedRunningTime="2025-12-04 17:55:13.85017701 +0000 UTC m=+975.805538076" watchObservedRunningTime="2025-12-04 17:55:13.853412807 +0000 UTC m=+975.808773853" Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.937660 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-2cpvp"] Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.938555 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-2cpvp" Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.942151 4733 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-xjksq" Dec 04 17:55:13 crc kubenswrapper[4733]: I1204 17:55:13.950891 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-2cpvp"] Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.055397 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvnrc\" (UniqueName: \"kubernetes.io/projected/1f3587c5-c7c0-40e2-855a-3b42ed4088c9-kube-api-access-nvnrc\") pod \"cert-manager-86cb77c54b-2cpvp\" (UID: \"1f3587c5-c7c0-40e2-855a-3b42ed4088c9\") " pod="cert-manager/cert-manager-86cb77c54b-2cpvp" Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.055528 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f3587c5-c7c0-40e2-855a-3b42ed4088c9-bound-sa-token\") pod \"cert-manager-86cb77c54b-2cpvp\" (UID: \"1f3587c5-c7c0-40e2-855a-3b42ed4088c9\") " pod="cert-manager/cert-manager-86cb77c54b-2cpvp" Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.157332 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvnrc\" (UniqueName: \"kubernetes.io/projected/1f3587c5-c7c0-40e2-855a-3b42ed4088c9-kube-api-access-nvnrc\") pod \"cert-manager-86cb77c54b-2cpvp\" (UID: \"1f3587c5-c7c0-40e2-855a-3b42ed4088c9\") " pod="cert-manager/cert-manager-86cb77c54b-2cpvp" Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.157418 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f3587c5-c7c0-40e2-855a-3b42ed4088c9-bound-sa-token\") pod \"cert-manager-86cb77c54b-2cpvp\" (UID: \"1f3587c5-c7c0-40e2-855a-3b42ed4088c9\") " pod="cert-manager/cert-manager-86cb77c54b-2cpvp" Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.180368 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f3587c5-c7c0-40e2-855a-3b42ed4088c9-bound-sa-token\") pod \"cert-manager-86cb77c54b-2cpvp\" (UID: \"1f3587c5-c7c0-40e2-855a-3b42ed4088c9\") " pod="cert-manager/cert-manager-86cb77c54b-2cpvp" Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.180546 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvnrc\" (UniqueName: \"kubernetes.io/projected/1f3587c5-c7c0-40e2-855a-3b42ed4088c9-kube-api-access-nvnrc\") pod \"cert-manager-86cb77c54b-2cpvp\" (UID: \"1f3587c5-c7c0-40e2-855a-3b42ed4088c9\") " pod="cert-manager/cert-manager-86cb77c54b-2cpvp" Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.299238 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-2cpvp" Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.563456 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-2cpvp"] Dec 04 17:55:14 crc kubenswrapper[4733]: W1204 17:55:14.573204 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f3587c5_c7c0_40e2_855a_3b42ed4088c9.slice/crio-ce0fe01c18a9e69d5e637e1170774446b82f3bab2ac86bf7abc1f595c91d8445 WatchSource:0}: Error finding container ce0fe01c18a9e69d5e637e1170774446b82f3bab2ac86bf7abc1f595c91d8445: Status 404 returned error can't find the container with id ce0fe01c18a9e69d5e637e1170774446b82f3bab2ac86bf7abc1f595c91d8445 Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.824512 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-2cpvp" event={"ID":"1f3587c5-c7c0-40e2-855a-3b42ed4088c9","Type":"ContainerStarted","Data":"83fc43dee56979e6a736482327768465b17243d95a83e60060d57ee1b3367370"} Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.824892 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-2cpvp" event={"ID":"1f3587c5-c7c0-40e2-855a-3b42ed4088c9","Type":"ContainerStarted","Data":"ce0fe01c18a9e69d5e637e1170774446b82f3bab2ac86bf7abc1f595c91d8445"} Dec 04 17:55:14 crc kubenswrapper[4733]: I1204 17:55:14.845438 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-2cpvp" podStartSLOduration=1.845419464 podStartE2EDuration="1.845419464s" podCreationTimestamp="2025-12-04 17:55:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:55:14.838784226 +0000 UTC m=+976.794145292" watchObservedRunningTime="2025-12-04 17:55:14.845419464 +0000 UTC m=+976.800780510" Dec 04 17:55:15 crc kubenswrapper[4733]: I1204 17:55:15.362598 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:55:15 crc kubenswrapper[4733]: I1204 17:55:15.362694 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:55:20 crc kubenswrapper[4733]: I1204 17:55:20.570954 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-k5b27" Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.519971 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-jtpvj"] Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.521403 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jtpvj" Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.524240 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.524931 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.525736 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-gqvfw" Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.552744 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqc6z\" (UniqueName: \"kubernetes.io/projected/bd890021-e67f-40ad-9fcd-fc301acba63f-kube-api-access-kqc6z\") pod \"openstack-operator-index-jtpvj\" (UID: \"bd890021-e67f-40ad-9fcd-fc301acba63f\") " pod="openstack-operators/openstack-operator-index-jtpvj" Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.580997 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jtpvj"] Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.653390 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqc6z\" (UniqueName: \"kubernetes.io/projected/bd890021-e67f-40ad-9fcd-fc301acba63f-kube-api-access-kqc6z\") pod \"openstack-operator-index-jtpvj\" (UID: \"bd890021-e67f-40ad-9fcd-fc301acba63f\") " pod="openstack-operators/openstack-operator-index-jtpvj" Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.675961 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqc6z\" (UniqueName: \"kubernetes.io/projected/bd890021-e67f-40ad-9fcd-fc301acba63f-kube-api-access-kqc6z\") pod \"openstack-operator-index-jtpvj\" (UID: \"bd890021-e67f-40ad-9fcd-fc301acba63f\") " pod="openstack-operators/openstack-operator-index-jtpvj" Dec 04 17:55:24 crc kubenswrapper[4733]: I1204 17:55:24.847093 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jtpvj" Dec 04 17:55:26 crc kubenswrapper[4733]: I1204 17:55:26.060081 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jtpvj"] Dec 04 17:55:26 crc kubenswrapper[4733]: I1204 17:55:26.949722 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jtpvj" event={"ID":"bd890021-e67f-40ad-9fcd-fc301acba63f","Type":"ContainerStarted","Data":"247e1d7eedf8d03db7c8497aa28cc7e85f083ee98b9f04a3440c8df12414c03e"} Dec 04 17:55:28 crc kubenswrapper[4733]: I1204 17:55:28.490625 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-jtpvj"] Dec 04 17:55:29 crc kubenswrapper[4733]: I1204 17:55:29.305875 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fmd92"] Dec 04 17:55:29 crc kubenswrapper[4733]: I1204 17:55:29.306672 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:29 crc kubenswrapper[4733]: I1204 17:55:29.318060 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22bbb\" (UniqueName: \"kubernetes.io/projected/0f74bf5a-9575-45d2-a08e-ac35a429ecdb-kube-api-access-22bbb\") pod \"openstack-operator-index-fmd92\" (UID: \"0f74bf5a-9575-45d2-a08e-ac35a429ecdb\") " pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:29 crc kubenswrapper[4733]: I1204 17:55:29.327683 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fmd92"] Dec 04 17:55:29 crc kubenswrapper[4733]: I1204 17:55:29.419559 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22bbb\" (UniqueName: \"kubernetes.io/projected/0f74bf5a-9575-45d2-a08e-ac35a429ecdb-kube-api-access-22bbb\") pod \"openstack-operator-index-fmd92\" (UID: \"0f74bf5a-9575-45d2-a08e-ac35a429ecdb\") " pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:29 crc kubenswrapper[4733]: I1204 17:55:29.443654 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22bbb\" (UniqueName: \"kubernetes.io/projected/0f74bf5a-9575-45d2-a08e-ac35a429ecdb-kube-api-access-22bbb\") pod \"openstack-operator-index-fmd92\" (UID: \"0f74bf5a-9575-45d2-a08e-ac35a429ecdb\") " pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:29 crc kubenswrapper[4733]: I1204 17:55:29.673612 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.179820 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fmd92"] Dec 04 17:55:30 crc kubenswrapper[4733]: W1204 17:55:30.189554 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f74bf5a_9575_45d2_a08e_ac35a429ecdb.slice/crio-6f17c671f1d3aa32b82cd0932ccb07c4d70174cf9f5404c448105faf7063bdc4 WatchSource:0}: Error finding container 6f17c671f1d3aa32b82cd0932ccb07c4d70174cf9f5404c448105faf7063bdc4: Status 404 returned error can't find the container with id 6f17c671f1d3aa32b82cd0932ccb07c4d70174cf9f5404c448105faf7063bdc4 Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.303291 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xhvtz"] Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.304447 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.321170 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xhvtz"] Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.341125 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-catalog-content\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.341183 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbkjz\" (UniqueName: \"kubernetes.io/projected/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-kube-api-access-lbkjz\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.341209 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-utilities\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.442041 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-catalog-content\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.442082 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbkjz\" (UniqueName: \"kubernetes.io/projected/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-kube-api-access-lbkjz\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.442108 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-utilities\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.442574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-utilities\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.442598 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-catalog-content\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.470876 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbkjz\" (UniqueName: \"kubernetes.io/projected/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-kube-api-access-lbkjz\") pod \"certified-operators-xhvtz\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.622822 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:30 crc kubenswrapper[4733]: I1204 17:55:30.987855 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fmd92" event={"ID":"0f74bf5a-9575-45d2-a08e-ac35a429ecdb","Type":"ContainerStarted","Data":"6f17c671f1d3aa32b82cd0932ccb07c4d70174cf9f5404c448105faf7063bdc4"} Dec 04 17:55:31 crc kubenswrapper[4733]: I1204 17:55:31.253727 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xhvtz"] Dec 04 17:55:31 crc kubenswrapper[4733]: W1204 17:55:31.259444 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33f8d1f5_0e0e_4a8a_a510_0fd25eac5033.slice/crio-70028c95c84506525d263fd113bf8a21405b3b5d520ebfc4e44f62c05f81a81b WatchSource:0}: Error finding container 70028c95c84506525d263fd113bf8a21405b3b5d520ebfc4e44f62c05f81a81b: Status 404 returned error can't find the container with id 70028c95c84506525d263fd113bf8a21405b3b5d520ebfc4e44f62c05f81a81b Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:31.999989 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jtpvj" event={"ID":"bd890021-e67f-40ad-9fcd-fc301acba63f","Type":"ContainerStarted","Data":"620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f"} Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.000111 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-jtpvj" podUID="bd890021-e67f-40ad-9fcd-fc301acba63f" containerName="registry-server" containerID="cri-o://620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f" gracePeriod=2 Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.006746 4733 generic.go:334] "Generic (PLEG): container finished" podID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerID="0abc69d946f313d5fe5626f09bef4b89bf4412401ecc2a84f113139809bd9cf6" exitCode=0 Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.006859 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhvtz" event={"ID":"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033","Type":"ContainerDied","Data":"0abc69d946f313d5fe5626f09bef4b89bf4412401ecc2a84f113139809bd9cf6"} Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.006967 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhvtz" event={"ID":"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033","Type":"ContainerStarted","Data":"70028c95c84506525d263fd113bf8a21405b3b5d520ebfc4e44f62c05f81a81b"} Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.011395 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fmd92" event={"ID":"0f74bf5a-9575-45d2-a08e-ac35a429ecdb","Type":"ContainerStarted","Data":"a5b5d268fbae723cd43b0869a3bd90fc17fdaeb73e516fe2f48ec1e9674aa10c"} Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.039342 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-jtpvj" podStartSLOduration=3.206981598 podStartE2EDuration="8.03931351s" podCreationTimestamp="2025-12-04 17:55:24 +0000 UTC" firstStartedPulling="2025-12-04 17:55:26.068958876 +0000 UTC m=+988.024319922" lastFinishedPulling="2025-12-04 17:55:30.901290798 +0000 UTC m=+992.856651834" observedRunningTime="2025-12-04 17:55:32.032143467 +0000 UTC m=+993.987504553" watchObservedRunningTime="2025-12-04 17:55:32.03931351 +0000 UTC m=+993.994674596" Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.054743 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fmd92" podStartSLOduration=2.342389456 podStartE2EDuration="3.054707524s" podCreationTimestamp="2025-12-04 17:55:29 +0000 UTC" firstStartedPulling="2025-12-04 17:55:30.192733381 +0000 UTC m=+992.148094437" lastFinishedPulling="2025-12-04 17:55:30.905051469 +0000 UTC m=+992.860412505" observedRunningTime="2025-12-04 17:55:32.05305908 +0000 UTC m=+994.008420166" watchObservedRunningTime="2025-12-04 17:55:32.054707524 +0000 UTC m=+994.010068640" Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.433143 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jtpvj" Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.570405 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqc6z\" (UniqueName: \"kubernetes.io/projected/bd890021-e67f-40ad-9fcd-fc301acba63f-kube-api-access-kqc6z\") pod \"bd890021-e67f-40ad-9fcd-fc301acba63f\" (UID: \"bd890021-e67f-40ad-9fcd-fc301acba63f\") " Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.578754 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd890021-e67f-40ad-9fcd-fc301acba63f-kube-api-access-kqc6z" (OuterVolumeSpecName: "kube-api-access-kqc6z") pod "bd890021-e67f-40ad-9fcd-fc301acba63f" (UID: "bd890021-e67f-40ad-9fcd-fc301acba63f"). InnerVolumeSpecName "kube-api-access-kqc6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:55:32 crc kubenswrapper[4733]: I1204 17:55:32.671764 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqc6z\" (UniqueName: \"kubernetes.io/projected/bd890021-e67f-40ad-9fcd-fc301acba63f-kube-api-access-kqc6z\") on node \"crc\" DevicePath \"\"" Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.033374 4733 generic.go:334] "Generic (PLEG): container finished" podID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerID="52f720f5791dfc973dad1a4a52deb0e11d41840dcdcc1f33d6870d6ed12e77e6" exitCode=0 Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.033451 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhvtz" event={"ID":"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033","Type":"ContainerDied","Data":"52f720f5791dfc973dad1a4a52deb0e11d41840dcdcc1f33d6870d6ed12e77e6"} Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.037284 4733 generic.go:334] "Generic (PLEG): container finished" podID="bd890021-e67f-40ad-9fcd-fc301acba63f" containerID="620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f" exitCode=0 Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.037474 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jtpvj" Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.037924 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jtpvj" event={"ID":"bd890021-e67f-40ad-9fcd-fc301acba63f","Type":"ContainerDied","Data":"620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f"} Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.037948 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jtpvj" event={"ID":"bd890021-e67f-40ad-9fcd-fc301acba63f","Type":"ContainerDied","Data":"247e1d7eedf8d03db7c8497aa28cc7e85f083ee98b9f04a3440c8df12414c03e"} Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.037970 4733 scope.go:117] "RemoveContainer" containerID="620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f" Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.079688 4733 scope.go:117] "RemoveContainer" containerID="620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f" Dec 04 17:55:33 crc kubenswrapper[4733]: E1204 17:55:33.080222 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f\": container with ID starting with 620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f not found: ID does not exist" containerID="620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f" Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.080275 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f"} err="failed to get container status \"620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f\": rpc error: code = NotFound desc = could not find container \"620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f\": container with ID starting with 620d6ca5bd5cb958598762772a9fcf9a0f28e3606d12b2a41bdfadf6bd8ee32f not found: ID does not exist" Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.085292 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-jtpvj"] Dec 04 17:55:33 crc kubenswrapper[4733]: I1204 17:55:33.093018 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-jtpvj"] Dec 04 17:55:34 crc kubenswrapper[4733]: I1204 17:55:34.060302 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhvtz" event={"ID":"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033","Type":"ContainerStarted","Data":"d2269d23776756431baa61533877cfc2261a88a16fa5933f141b321721f066a8"} Dec 04 17:55:34 crc kubenswrapper[4733]: I1204 17:55:34.091828 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xhvtz" podStartSLOduration=2.686729433 podStartE2EDuration="4.091757234s" podCreationTimestamp="2025-12-04 17:55:30 +0000 UTC" firstStartedPulling="2025-12-04 17:55:32.009377115 +0000 UTC m=+993.964738201" lastFinishedPulling="2025-12-04 17:55:33.414404926 +0000 UTC m=+995.369766002" observedRunningTime="2025-12-04 17:55:34.085634809 +0000 UTC m=+996.040995905" watchObservedRunningTime="2025-12-04 17:55:34.091757234 +0000 UTC m=+996.047118320" Dec 04 17:55:34 crc kubenswrapper[4733]: I1204 17:55:34.348684 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd890021-e67f-40ad-9fcd-fc301acba63f" path="/var/lib/kubelet/pods/bd890021-e67f-40ad-9fcd-fc301acba63f/volumes" Dec 04 17:55:39 crc kubenswrapper[4733]: I1204 17:55:39.674554 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:39 crc kubenswrapper[4733]: I1204 17:55:39.675001 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:39 crc kubenswrapper[4733]: I1204 17:55:39.732511 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:40 crc kubenswrapper[4733]: I1204 17:55:40.146684 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-fmd92" Dec 04 17:55:40 crc kubenswrapper[4733]: I1204 17:55:40.623578 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:40 crc kubenswrapper[4733]: I1204 17:55:40.624173 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:40 crc kubenswrapper[4733]: I1204 17:55:40.696363 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:41 crc kubenswrapper[4733]: I1204 17:55:41.190585 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:42 crc kubenswrapper[4733]: I1204 17:55:42.094422 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xhvtz"] Dec 04 17:55:43 crc kubenswrapper[4733]: I1204 17:55:43.134611 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xhvtz" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerName="registry-server" containerID="cri-o://d2269d23776756431baa61533877cfc2261a88a16fa5933f141b321721f066a8" gracePeriod=2 Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.149537 4733 generic.go:334] "Generic (PLEG): container finished" podID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerID="d2269d23776756431baa61533877cfc2261a88a16fa5933f141b321721f066a8" exitCode=0 Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.149586 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhvtz" event={"ID":"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033","Type":"ContainerDied","Data":"d2269d23776756431baa61533877cfc2261a88a16fa5933f141b321721f066a8"} Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.687964 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.799044 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-catalog-content\") pod \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.799165 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbkjz\" (UniqueName: \"kubernetes.io/projected/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-kube-api-access-lbkjz\") pod \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.799223 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-utilities\") pod \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\" (UID: \"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033\") " Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.800534 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-utilities" (OuterVolumeSpecName: "utilities") pod "33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" (UID: "33f8d1f5-0e0e-4a8a-a510-0fd25eac5033"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.811229 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-kube-api-access-lbkjz" (OuterVolumeSpecName: "kube-api-access-lbkjz") pod "33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" (UID: "33f8d1f5-0e0e-4a8a-a510-0fd25eac5033"). InnerVolumeSpecName "kube-api-access-lbkjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.856919 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" (UID: "33f8d1f5-0e0e-4a8a-a510-0fd25eac5033"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.902292 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbkjz\" (UniqueName: \"kubernetes.io/projected/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-kube-api-access-lbkjz\") on node \"crc\" DevicePath \"\"" Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.902355 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 17:55:44 crc kubenswrapper[4733]: I1204 17:55:44.902368 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.161336 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhvtz" event={"ID":"33f8d1f5-0e0e-4a8a-a510-0fd25eac5033","Type":"ContainerDied","Data":"70028c95c84506525d263fd113bf8a21405b3b5d520ebfc4e44f62c05f81a81b"} Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.161420 4733 scope.go:117] "RemoveContainer" containerID="d2269d23776756431baa61533877cfc2261a88a16fa5933f141b321721f066a8" Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.161423 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhvtz" Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.185544 4733 scope.go:117] "RemoveContainer" containerID="52f720f5791dfc973dad1a4a52deb0e11d41840dcdcc1f33d6870d6ed12e77e6" Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.207223 4733 scope.go:117] "RemoveContainer" containerID="0abc69d946f313d5fe5626f09bef4b89bf4412401ecc2a84f113139809bd9cf6" Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.211329 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xhvtz"] Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.217005 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xhvtz"] Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.362246 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:55:45 crc kubenswrapper[4733]: I1204 17:55:45.362336 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:55:46 crc kubenswrapper[4733]: I1204 17:55:46.344432 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" path="/var/lib/kubelet/pods/33f8d1f5-0e0e-4a8a-a510-0fd25eac5033/volumes" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.765901 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx"] Dec 04 17:55:53 crc kubenswrapper[4733]: E1204 17:55:53.766651 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd890021-e67f-40ad-9fcd-fc301acba63f" containerName="registry-server" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.766674 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd890021-e67f-40ad-9fcd-fc301acba63f" containerName="registry-server" Dec 04 17:55:53 crc kubenswrapper[4733]: E1204 17:55:53.766696 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerName="registry-server" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.766714 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerName="registry-server" Dec 04 17:55:53 crc kubenswrapper[4733]: E1204 17:55:53.766734 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerName="extract-utilities" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.766748 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerName="extract-utilities" Dec 04 17:55:53 crc kubenswrapper[4733]: E1204 17:55:53.766787 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerName="extract-content" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.766840 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerName="extract-content" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.767063 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="33f8d1f5-0e0e-4a8a-a510-0fd25eac5033" containerName="registry-server" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.767098 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd890021-e67f-40ad-9fcd-fc301acba63f" containerName="registry-server" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.768757 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.775443 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qfzzl" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.777336 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx"] Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.933741 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgg95\" (UniqueName: \"kubernetes.io/projected/5e268882-1f3e-40a2-a37c-2ad332106598-kube-api-access-xgg95\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.934040 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:53 crc kubenswrapper[4733]: I1204 17:55:53.934286 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:54 crc kubenswrapper[4733]: I1204 17:55:54.035678 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgg95\" (UniqueName: \"kubernetes.io/projected/5e268882-1f3e-40a2-a37c-2ad332106598-kube-api-access-xgg95\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:54 crc kubenswrapper[4733]: I1204 17:55:54.035836 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:54 crc kubenswrapper[4733]: I1204 17:55:54.035874 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:54 crc kubenswrapper[4733]: I1204 17:55:54.036376 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:54 crc kubenswrapper[4733]: I1204 17:55:54.036912 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:54 crc kubenswrapper[4733]: I1204 17:55:54.074850 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgg95\" (UniqueName: \"kubernetes.io/projected/5e268882-1f3e-40a2-a37c-2ad332106598-kube-api-access-xgg95\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:54 crc kubenswrapper[4733]: I1204 17:55:54.100643 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:54 crc kubenswrapper[4733]: I1204 17:55:54.583987 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx"] Dec 04 17:55:55 crc kubenswrapper[4733]: I1204 17:55:55.267749 4733 generic.go:334] "Generic (PLEG): container finished" podID="5e268882-1f3e-40a2-a37c-2ad332106598" containerID="a4f2f10e175f1fe4a4667a3020e5b36cb1fc3cbbf675fdc7e3c00e2f4271d388" exitCode=0 Dec 04 17:55:55 crc kubenswrapper[4733]: I1204 17:55:55.267848 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" event={"ID":"5e268882-1f3e-40a2-a37c-2ad332106598","Type":"ContainerDied","Data":"a4f2f10e175f1fe4a4667a3020e5b36cb1fc3cbbf675fdc7e3c00e2f4271d388"} Dec 04 17:55:55 crc kubenswrapper[4733]: I1204 17:55:55.267885 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" event={"ID":"5e268882-1f3e-40a2-a37c-2ad332106598","Type":"ContainerStarted","Data":"389eaa379b21e2172854f9d2e678c23edfc8e3e98c1df93990526706ff8bb806"} Dec 04 17:55:57 crc kubenswrapper[4733]: I1204 17:55:57.290050 4733 generic.go:334] "Generic (PLEG): container finished" podID="5e268882-1f3e-40a2-a37c-2ad332106598" containerID="ac43211d74437e758c526e9e1a278b7bb1e0e119f28630948866dffd1bb222bb" exitCode=0 Dec 04 17:55:57 crc kubenswrapper[4733]: I1204 17:55:57.290156 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" event={"ID":"5e268882-1f3e-40a2-a37c-2ad332106598","Type":"ContainerDied","Data":"ac43211d74437e758c526e9e1a278b7bb1e0e119f28630948866dffd1bb222bb"} Dec 04 17:55:58 crc kubenswrapper[4733]: I1204 17:55:58.305108 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" event={"ID":"5e268882-1f3e-40a2-a37c-2ad332106598","Type":"ContainerDied","Data":"cbc8f71526f75e2be25defaa393fa62f98fb69617fe79e30c89862373ec34531"} Dec 04 17:55:58 crc kubenswrapper[4733]: I1204 17:55:58.305056 4733 generic.go:334] "Generic (PLEG): container finished" podID="5e268882-1f3e-40a2-a37c-2ad332106598" containerID="cbc8f71526f75e2be25defaa393fa62f98fb69617fe79e30c89862373ec34531" exitCode=0 Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.713053 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.858361 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-bundle\") pod \"5e268882-1f3e-40a2-a37c-2ad332106598\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.858432 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-util\") pod \"5e268882-1f3e-40a2-a37c-2ad332106598\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.858522 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgg95\" (UniqueName: \"kubernetes.io/projected/5e268882-1f3e-40a2-a37c-2ad332106598-kube-api-access-xgg95\") pod \"5e268882-1f3e-40a2-a37c-2ad332106598\" (UID: \"5e268882-1f3e-40a2-a37c-2ad332106598\") " Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.859309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-bundle" (OuterVolumeSpecName: "bundle") pod "5e268882-1f3e-40a2-a37c-2ad332106598" (UID: "5e268882-1f3e-40a2-a37c-2ad332106598"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.865641 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e268882-1f3e-40a2-a37c-2ad332106598-kube-api-access-xgg95" (OuterVolumeSpecName: "kube-api-access-xgg95") pod "5e268882-1f3e-40a2-a37c-2ad332106598" (UID: "5e268882-1f3e-40a2-a37c-2ad332106598"). InnerVolumeSpecName "kube-api-access-xgg95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.889119 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-util" (OuterVolumeSpecName: "util") pod "5e268882-1f3e-40a2-a37c-2ad332106598" (UID: "5e268882-1f3e-40a2-a37c-2ad332106598"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.960331 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgg95\" (UniqueName: \"kubernetes.io/projected/5e268882-1f3e-40a2-a37c-2ad332106598-kube-api-access-xgg95\") on node \"crc\" DevicePath \"\"" Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.960372 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:55:59 crc kubenswrapper[4733]: I1204 17:55:59.960388 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e268882-1f3e-40a2-a37c-2ad332106598-util\") on node \"crc\" DevicePath \"\"" Dec 04 17:56:00 crc kubenswrapper[4733]: I1204 17:56:00.326029 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" event={"ID":"5e268882-1f3e-40a2-a37c-2ad332106598","Type":"ContainerDied","Data":"389eaa379b21e2172854f9d2e678c23edfc8e3e98c1df93990526706ff8bb806"} Dec 04 17:56:00 crc kubenswrapper[4733]: I1204 17:56:00.326096 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="389eaa379b21e2172854f9d2e678c23edfc8e3e98c1df93990526706ff8bb806" Dec 04 17:56:00 crc kubenswrapper[4733]: I1204 17:56:00.326147 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.138931 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg"] Dec 04 17:56:02 crc kubenswrapper[4733]: E1204 17:56:02.139686 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e268882-1f3e-40a2-a37c-2ad332106598" containerName="util" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.139701 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e268882-1f3e-40a2-a37c-2ad332106598" containerName="util" Dec 04 17:56:02 crc kubenswrapper[4733]: E1204 17:56:02.139721 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e268882-1f3e-40a2-a37c-2ad332106598" containerName="pull" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.139727 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e268882-1f3e-40a2-a37c-2ad332106598" containerName="pull" Dec 04 17:56:02 crc kubenswrapper[4733]: E1204 17:56:02.139752 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e268882-1f3e-40a2-a37c-2ad332106598" containerName="extract" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.139758 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e268882-1f3e-40a2-a37c-2ad332106598" containerName="extract" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.139889 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e268882-1f3e-40a2-a37c-2ad332106598" containerName="extract" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.140431 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.143674 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-xv4vm" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.160255 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg"] Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.291193 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ql29\" (UniqueName: \"kubernetes.io/projected/93a96c03-299d-47fa-9873-2c890b731890-kube-api-access-2ql29\") pod \"openstack-operator-controller-operator-55b6fb9447-47stg\" (UID: \"93a96c03-299d-47fa-9873-2c890b731890\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.392843 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ql29\" (UniqueName: \"kubernetes.io/projected/93a96c03-299d-47fa-9873-2c890b731890-kube-api-access-2ql29\") pod \"openstack-operator-controller-operator-55b6fb9447-47stg\" (UID: \"93a96c03-299d-47fa-9873-2c890b731890\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.413811 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ql29\" (UniqueName: \"kubernetes.io/projected/93a96c03-299d-47fa-9873-2c890b731890-kube-api-access-2ql29\") pod \"openstack-operator-controller-operator-55b6fb9447-47stg\" (UID: \"93a96c03-299d-47fa-9873-2c890b731890\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.455624 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" Dec 04 17:56:02 crc kubenswrapper[4733]: I1204 17:56:02.992467 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg"] Dec 04 17:56:03 crc kubenswrapper[4733]: I1204 17:56:03.346456 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" event={"ID":"93a96c03-299d-47fa-9873-2c890b731890","Type":"ContainerStarted","Data":"9b69adf7fa98bf1328cc4719edbeca5b72b16a9d23fa89a7bc594e4a3b36de7c"} Dec 04 17:56:07 crc kubenswrapper[4733]: I1204 17:56:07.368124 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" event={"ID":"93a96c03-299d-47fa-9873-2c890b731890","Type":"ContainerStarted","Data":"042e8526bd39493435c228f2f380bb796f32537f42e876a23e4615569fb22301"} Dec 04 17:56:07 crc kubenswrapper[4733]: I1204 17:56:07.368542 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" Dec 04 17:56:07 crc kubenswrapper[4733]: I1204 17:56:07.399947 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" podStartSLOduration=1.367103893 podStartE2EDuration="5.399929274s" podCreationTimestamp="2025-12-04 17:56:02 +0000 UTC" firstStartedPulling="2025-12-04 17:56:03.009391678 +0000 UTC m=+1024.964752744" lastFinishedPulling="2025-12-04 17:56:07.042217079 +0000 UTC m=+1028.997578125" observedRunningTime="2025-12-04 17:56:07.39418482 +0000 UTC m=+1029.349545876" watchObservedRunningTime="2025-12-04 17:56:07.399929274 +0000 UTC m=+1029.355290320" Dec 04 17:56:12 crc kubenswrapper[4733]: I1204 17:56:12.459406 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-47stg" Dec 04 17:56:15 crc kubenswrapper[4733]: I1204 17:56:15.362646 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:56:15 crc kubenswrapper[4733]: I1204 17:56:15.363206 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:56:15 crc kubenswrapper[4733]: I1204 17:56:15.363295 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:56:15 crc kubenswrapper[4733]: I1204 17:56:15.364340 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"245d5f7c913eadd0235bcdf1e7f003bb02e52afa3a7ef2fd2c9a9a3603961fd8"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 17:56:15 crc kubenswrapper[4733]: I1204 17:56:15.364446 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://245d5f7c913eadd0235bcdf1e7f003bb02e52afa3a7ef2fd2c9a9a3603961fd8" gracePeriod=600 Dec 04 17:56:16 crc kubenswrapper[4733]: I1204 17:56:16.437664 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="245d5f7c913eadd0235bcdf1e7f003bb02e52afa3a7ef2fd2c9a9a3603961fd8" exitCode=0 Dec 04 17:56:16 crc kubenswrapper[4733]: I1204 17:56:16.437742 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"245d5f7c913eadd0235bcdf1e7f003bb02e52afa3a7ef2fd2c9a9a3603961fd8"} Dec 04 17:56:16 crc kubenswrapper[4733]: I1204 17:56:16.438335 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"f00595a749ca89ac07966111bfcf1cdeb8fcd3b039701228e808a5c55fac2e38"} Dec 04 17:56:16 crc kubenswrapper[4733]: I1204 17:56:16.438366 4733 scope.go:117] "RemoveContainer" containerID="ee5a68aff544c2577446d346027585de3353ade4bb096661850d64dabfc79180" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.655296 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.657052 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.658505 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-lpgxb" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.664560 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.666238 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.669043 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5ppmc" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.669581 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.671143 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.672712 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-n86m8" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.674618 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.681272 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.695116 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.707319 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.708202 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.715309 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-hwnnx" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.718675 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.721093 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.724081 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-lr8z4" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.736184 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.760962 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnspv\" (UniqueName: \"kubernetes.io/projected/500a0dd4-904a-449e-956a-013bed8c66ca-kube-api-access-pnspv\") pod \"cinder-operator-controller-manager-859b6ccc6-8t88q\" (UID: \"500a0dd4-904a-449e-956a-013bed8c66ca\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.761024 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg2h2\" (UniqueName: \"kubernetes.io/projected/056b1e59-a544-4274-ae2e-b3d35ea1ab06-kube-api-access-dg2h2\") pod \"designate-operator-controller-manager-78b4bc895b-cs6cz\" (UID: \"056b1e59-a544-4274-ae2e-b3d35ea1ab06\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.761049 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvx2f\" (UniqueName: \"kubernetes.io/projected/9e110407-2b5b-4f7a-b815-77203c99b56b-kube-api-access-cvx2f\") pod \"barbican-operator-controller-manager-7d9dfd778-m45ps\" (UID: \"9e110407-2b5b-4f7a-b815-77203c99b56b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.761080 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6mhz\" (UniqueName: \"kubernetes.io/projected/f4fddbf6-0bc5-4b36-9a8a-7073e875c540-kube-api-access-w6mhz\") pod \"heat-operator-controller-manager-5f64f6f8bb-25jzc\" (UID: \"f4fddbf6-0bc5-4b36-9a8a-7073e875c540\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.761110 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g5gs\" (UniqueName: \"kubernetes.io/projected/fbad1f3f-55ee-4477-a8a4-c146ee09ba51-kube-api-access-7g5gs\") pod \"glance-operator-controller-manager-77987cd8cd-bjzvv\" (UID: \"fbad1f3f-55ee-4477-a8a4-c146ee09ba51\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.772339 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.784509 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.786100 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.792243 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-xwb7k" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.816065 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-8skx7"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.817335 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.820124 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-m695k" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.820328 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.845855 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.846851 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.850944 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-2xvtj" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.861776 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnspv\" (UniqueName: \"kubernetes.io/projected/500a0dd4-904a-449e-956a-013bed8c66ca-kube-api-access-pnspv\") pod \"cinder-operator-controller-manager-859b6ccc6-8t88q\" (UID: \"500a0dd4-904a-449e-956a-013bed8c66ca\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.861893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg2h2\" (UniqueName: \"kubernetes.io/projected/056b1e59-a544-4274-ae2e-b3d35ea1ab06-kube-api-access-dg2h2\") pod \"designate-operator-controller-manager-78b4bc895b-cs6cz\" (UID: \"056b1e59-a544-4274-ae2e-b3d35ea1ab06\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.861922 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lblwn\" (UniqueName: \"kubernetes.io/projected/2191192e-388c-4535-be27-5b4b76aea175-kube-api-access-lblwn\") pod \"horizon-operator-controller-manager-68c6d99b8f-2pm4z\" (UID: \"2191192e-388c-4535-be27-5b4b76aea175\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.861942 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvx2f\" (UniqueName: \"kubernetes.io/projected/9e110407-2b5b-4f7a-b815-77203c99b56b-kube-api-access-cvx2f\") pod \"barbican-operator-controller-manager-7d9dfd778-m45ps\" (UID: \"9e110407-2b5b-4f7a-b815-77203c99b56b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.861972 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tlrd\" (UniqueName: \"kubernetes.io/projected/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-kube-api-access-9tlrd\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.861991 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6mhz\" (UniqueName: \"kubernetes.io/projected/f4fddbf6-0bc5-4b36-9a8a-7073e875c540-kube-api-access-w6mhz\") pod \"heat-operator-controller-manager-5f64f6f8bb-25jzc\" (UID: \"f4fddbf6-0bc5-4b36-9a8a-7073e875c540\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.862013 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.862030 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g5gs\" (UniqueName: \"kubernetes.io/projected/fbad1f3f-55ee-4477-a8a4-c146ee09ba51-kube-api-access-7g5gs\") pod \"glance-operator-controller-manager-77987cd8cd-bjzvv\" (UID: \"fbad1f3f-55ee-4477-a8a4-c146ee09ba51\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.869024 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-8skx7"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.878466 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.879515 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.884845 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.891108 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.897283 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-mvp4v" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.897988 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g5gs\" (UniqueName: \"kubernetes.io/projected/fbad1f3f-55ee-4477-a8a4-c146ee09ba51-kube-api-access-7g5gs\") pod \"glance-operator-controller-manager-77987cd8cd-bjzvv\" (UID: \"fbad1f3f-55ee-4477-a8a4-c146ee09ba51\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.899979 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.901061 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.901466 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnspv\" (UniqueName: \"kubernetes.io/projected/500a0dd4-904a-449e-956a-013bed8c66ca-kube-api-access-pnspv\") pod \"cinder-operator-controller-manager-859b6ccc6-8t88q\" (UID: \"500a0dd4-904a-449e-956a-013bed8c66ca\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.905608 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-zk8dk" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.905750 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.906699 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.906717 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6mhz\" (UniqueName: \"kubernetes.io/projected/f4fddbf6-0bc5-4b36-9a8a-7073e875c540-kube-api-access-w6mhz\") pod \"heat-operator-controller-manager-5f64f6f8bb-25jzc\" (UID: \"f4fddbf6-0bc5-4b36-9a8a-7073e875c540\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.907373 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg2h2\" (UniqueName: \"kubernetes.io/projected/056b1e59-a544-4274-ae2e-b3d35ea1ab06-kube-api-access-dg2h2\") pod \"designate-operator-controller-manager-78b4bc895b-cs6cz\" (UID: \"056b1e59-a544-4274-ae2e-b3d35ea1ab06\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.910302 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvx2f\" (UniqueName: \"kubernetes.io/projected/9e110407-2b5b-4f7a-b815-77203c99b56b-kube-api-access-cvx2f\") pod \"barbican-operator-controller-manager-7d9dfd778-m45ps\" (UID: \"9e110407-2b5b-4f7a-b815-77203c99b56b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.911990 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-rdftr" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.926771 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.927848 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.933350 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-v4cdq" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.937541 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.945312 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.959639 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.962887 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzkp7\" (UniqueName: \"kubernetes.io/projected/50dfe63c-b3b2-46c5-ad8d-91c64e3ef698-kube-api-access-tzkp7\") pod \"keystone-operator-controller-manager-7765d96ddf-qpssp\" (UID: \"50dfe63c-b3b2-46c5-ad8d-91c64e3ef698\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.962934 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w2wn\" (UniqueName: \"kubernetes.io/projected/143e7c55-1ecb-4efa-8992-332718ae5b61-kube-api-access-9w2wn\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vqnt5\" (UID: \"143e7c55-1ecb-4efa-8992-332718ae5b61\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.962989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhtvr\" (UniqueName: \"kubernetes.io/projected/e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5-kube-api-access-xhtvr\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-txjrm\" (UID: \"e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.963067 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lblwn\" (UniqueName: \"kubernetes.io/projected/2191192e-388c-4535-be27-5b4b76aea175-kube-api-access-lblwn\") pod \"horizon-operator-controller-manager-68c6d99b8f-2pm4z\" (UID: \"2191192e-388c-4535-be27-5b4b76aea175\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.963088 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njgs4\" (UniqueName: \"kubernetes.io/projected/7f29393d-5068-4870-86ab-237c52f6d1f9-kube-api-access-njgs4\") pod \"manila-operator-controller-manager-7c79b5df47-8r2s2\" (UID: \"7f29393d-5068-4870-86ab-237c52f6d1f9\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.963116 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tlrd\" (UniqueName: \"kubernetes.io/projected/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-kube-api-access-9tlrd\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.963167 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.963194 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2djhv\" (UniqueName: \"kubernetes.io/projected/da343538-3e4a-49a0-949d-2622d62ed6a2-kube-api-access-2djhv\") pod \"ironic-operator-controller-manager-6c548fd776-v8d4h\" (UID: \"da343538-3e4a-49a0-949d-2622d62ed6a2\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" Dec 04 17:56:46 crc kubenswrapper[4733]: E1204 17:56:46.963598 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:46 crc kubenswrapper[4733]: E1204 17:56:46.963643 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert podName:504dcc38-54b7-430f-8c32-e11e5bbbf5dc nodeName:}" failed. No retries permitted until 2025-12-04 17:56:47.463629957 +0000 UTC m=+1069.418991003 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert") pod "infra-operator-controller-manager-57548d458d-8skx7" (UID: "504dcc38-54b7-430f-8c32-e11e5bbbf5dc") : secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.978853 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.980632 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm"] Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.991024 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lblwn\" (UniqueName: \"kubernetes.io/projected/2191192e-388c-4535-be27-5b4b76aea175-kube-api-access-lblwn\") pod \"horizon-operator-controller-manager-68c6d99b8f-2pm4z\" (UID: \"2191192e-388c-4535-be27-5b4b76aea175\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.993865 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" Dec 04 17:56:46 crc kubenswrapper[4733]: I1204 17:56:46.996720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tlrd\" (UniqueName: \"kubernetes.io/projected/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-kube-api-access-9tlrd\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.003973 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.011539 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.019641 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.028170 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.028405 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.029305 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.032438 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-nf8dv" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.032335 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-94fgj" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.057187 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.061847 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.062944 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.063042 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.063913 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhtvr\" (UniqueName: \"kubernetes.io/projected/e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5-kube-api-access-xhtvr\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-txjrm\" (UID: \"e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.063942 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxln6\" (UniqueName: \"kubernetes.io/projected/dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8-kube-api-access-sxln6\") pod \"nova-operator-controller-manager-697bc559fc-5rln8\" (UID: \"dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.063968 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njgs4\" (UniqueName: \"kubernetes.io/projected/7f29393d-5068-4870-86ab-237c52f6d1f9-kube-api-access-njgs4\") pod \"manila-operator-controller-manager-7c79b5df47-8r2s2\" (UID: \"7f29393d-5068-4870-86ab-237c52f6d1f9\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.064003 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tvz4\" (UniqueName: \"kubernetes.io/projected/889320d7-9fc9-43b9-be8b-d5d8ceeb966c-kube-api-access-9tvz4\") pod \"octavia-operator-controller-manager-998648c74-k5gsb\" (UID: \"889320d7-9fc9-43b9-be8b-d5d8ceeb966c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.064048 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2djhv\" (UniqueName: \"kubernetes.io/projected/da343538-3e4a-49a0-949d-2622d62ed6a2-kube-api-access-2djhv\") pod \"ironic-operator-controller-manager-6c548fd776-v8d4h\" (UID: \"da343538-3e4a-49a0-949d-2622d62ed6a2\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.064083 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzkp7\" (UniqueName: \"kubernetes.io/projected/50dfe63c-b3b2-46c5-ad8d-91c64e3ef698-kube-api-access-tzkp7\") pod \"keystone-operator-controller-manager-7765d96ddf-qpssp\" (UID: \"50dfe63c-b3b2-46c5-ad8d-91c64e3ef698\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.064110 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w2wn\" (UniqueName: \"kubernetes.io/projected/143e7c55-1ecb-4efa-8992-332718ae5b61-kube-api-access-9w2wn\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vqnt5\" (UID: \"143e7c55-1ecb-4efa-8992-332718ae5b61\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.071120 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-pfn2q" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.071537 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.085648 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.102582 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.111944 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njgs4\" (UniqueName: \"kubernetes.io/projected/7f29393d-5068-4870-86ab-237c52f6d1f9-kube-api-access-njgs4\") pod \"manila-operator-controller-manager-7c79b5df47-8r2s2\" (UID: \"7f29393d-5068-4870-86ab-237c52f6d1f9\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.114445 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzkp7\" (UniqueName: \"kubernetes.io/projected/50dfe63c-b3b2-46c5-ad8d-91c64e3ef698-kube-api-access-tzkp7\") pod \"keystone-operator-controller-manager-7765d96ddf-qpssp\" (UID: \"50dfe63c-b3b2-46c5-ad8d-91c64e3ef698\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.114569 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w2wn\" (UniqueName: \"kubernetes.io/projected/143e7c55-1ecb-4efa-8992-332718ae5b61-kube-api-access-9w2wn\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vqnt5\" (UID: \"143e7c55-1ecb-4efa-8992-332718ae5b61\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.114939 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2djhv\" (UniqueName: \"kubernetes.io/projected/da343538-3e4a-49a0-949d-2622d62ed6a2-kube-api-access-2djhv\") pod \"ironic-operator-controller-manager-6c548fd776-v8d4h\" (UID: \"da343538-3e4a-49a0-949d-2622d62ed6a2\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.115742 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhtvr\" (UniqueName: \"kubernetes.io/projected/e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5-kube-api-access-xhtvr\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-txjrm\" (UID: \"e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.139180 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.166983 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxln6\" (UniqueName: \"kubernetes.io/projected/dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8-kube-api-access-sxln6\") pod \"nova-operator-controller-manager-697bc559fc-5rln8\" (UID: \"dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.167332 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.167387 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tvz4\" (UniqueName: \"kubernetes.io/projected/889320d7-9fc9-43b9-be8b-d5d8ceeb966c-kube-api-access-9tvz4\") pod \"octavia-operator-controller-manager-998648c74-k5gsb\" (UID: \"889320d7-9fc9-43b9-be8b-d5d8ceeb966c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.167481 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgwbp\" (UniqueName: \"kubernetes.io/projected/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-kube-api-access-pgwbp\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.168546 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.180997 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.182635 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.190668 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tvz4\" (UniqueName: \"kubernetes.io/projected/889320d7-9fc9-43b9-be8b-d5d8ceeb966c-kube-api-access-9tvz4\") pod \"octavia-operator-controller-manager-998648c74-k5gsb\" (UID: \"889320d7-9fc9-43b9-be8b-d5d8ceeb966c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.194012 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-mrhs2" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.211705 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxln6\" (UniqueName: \"kubernetes.io/projected/dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8-kube-api-access-sxln6\") pod \"nova-operator-controller-manager-697bc559fc-5rln8\" (UID: \"dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.225265 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.226408 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.229040 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-2fshd" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.240546 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.259137 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.272865 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgwbp\" (UniqueName: \"kubernetes.io/projected/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-kube-api-access-pgwbp\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.272927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.272960 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnctp\" (UniqueName: \"kubernetes.io/projected/6c895222-dbc2-4490-972c-b7e9dad9e839-kube-api-access-jnctp\") pod \"placement-operator-controller-manager-78f8948974-7qmwh\" (UID: \"6c895222-dbc2-4490-972c-b7e9dad9e839\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.273025 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b7s8\" (UniqueName: \"kubernetes.io/projected/d0217980-ef7c-4b3a-b5cf-9887545b7a13-kube-api-access-7b7s8\") pod \"ovn-operator-controller-manager-b6456fdb6-v8s48\" (UID: \"d0217980-ef7c-4b3a-b5cf-9887545b7a13\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.273328 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.273363 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert podName:38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee nodeName:}" failed. No retries permitted until 2025-12-04 17:56:47.773350924 +0000 UTC m=+1069.728711960 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f59cdsz" (UID: "38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.273675 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.312166 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgwbp\" (UniqueName: \"kubernetes.io/projected/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-kube-api-access-pgwbp\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.312228 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.313233 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.318853 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.319182 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-rsphh" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.325432 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.380110 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b7s8\" (UniqueName: \"kubernetes.io/projected/d0217980-ef7c-4b3a-b5cf-9887545b7a13-kube-api-access-7b7s8\") pod \"ovn-operator-controller-manager-b6456fdb6-v8s48\" (UID: \"d0217980-ef7c-4b3a-b5cf-9887545b7a13\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.380316 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddmnd\" (UniqueName: \"kubernetes.io/projected/89feaf19-9df3-415a-8bbf-57168c3b89ba-kube-api-access-ddmnd\") pod \"swift-operator-controller-manager-5f8c65bbfc-7rb9w\" (UID: \"89feaf19-9df3-415a-8bbf-57168c3b89ba\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.380391 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnctp\" (UniqueName: \"kubernetes.io/projected/6c895222-dbc2-4490-972c-b7e9dad9e839-kube-api-access-jnctp\") pod \"placement-operator-controller-manager-78f8948974-7qmwh\" (UID: \"6c895222-dbc2-4490-972c-b7e9dad9e839\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.382002 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.387112 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.409253 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.424223 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.425819 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.426907 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b7s8\" (UniqueName: \"kubernetes.io/projected/d0217980-ef7c-4b3a-b5cf-9887545b7a13-kube-api-access-7b7s8\") pod \"ovn-operator-controller-manager-b6456fdb6-v8s48\" (UID: \"d0217980-ef7c-4b3a-b5cf-9887545b7a13\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.427358 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.427571 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnctp\" (UniqueName: \"kubernetes.io/projected/6c895222-dbc2-4490-972c-b7e9dad9e839-kube-api-access-jnctp\") pod \"placement-operator-controller-manager-78f8948974-7qmwh\" (UID: \"6c895222-dbc2-4490-972c-b7e9dad9e839\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.428406 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-79vc5" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.457735 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.482007 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch548\" (UniqueName: \"kubernetes.io/projected/0b437634-e4b9-4e70-b501-2aee9670efea-kube-api-access-ch548\") pod \"telemetry-operator-controller-manager-76cc84c6bb-zw57g\" (UID: \"0b437634-e4b9-4e70-b501-2aee9670efea\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.482086 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddmnd\" (UniqueName: \"kubernetes.io/projected/89feaf19-9df3-415a-8bbf-57168c3b89ba-kube-api-access-ddmnd\") pod \"swift-operator-controller-manager-5f8c65bbfc-7rb9w\" (UID: \"89feaf19-9df3-415a-8bbf-57168c3b89ba\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.482142 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.482283 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.482365 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert podName:504dcc38-54b7-430f-8c32-e11e5bbbf5dc nodeName:}" failed. No retries permitted until 2025-12-04 17:56:48.482347221 +0000 UTC m=+1070.437708267 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert") pod "infra-operator-controller-manager-57548d458d-8skx7" (UID: "504dcc38-54b7-430f-8c32-e11e5bbbf5dc") : secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.493930 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.500751 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.503904 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-j8xkj" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.512202 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddmnd\" (UniqueName: \"kubernetes.io/projected/89feaf19-9df3-415a-8bbf-57168c3b89ba-kube-api-access-ddmnd\") pod \"swift-operator-controller-manager-5f8c65bbfc-7rb9w\" (UID: \"89feaf19-9df3-415a-8bbf-57168c3b89ba\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.525927 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.528739 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.561667 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.565482 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.566805 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.570633 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-zr8z4" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.571131 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.587529 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch548\" (UniqueName: \"kubernetes.io/projected/0b437634-e4b9-4e70-b501-2aee9670efea-kube-api-access-ch548\") pod \"telemetry-operator-controller-manager-76cc84c6bb-zw57g\" (UID: \"0b437634-e4b9-4e70-b501-2aee9670efea\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.597847 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.598696 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.606100 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.606277 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.606721 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-4cxqg" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.612151 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.621384 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.623829 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.630357 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8s4rc" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.636405 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch548\" (UniqueName: \"kubernetes.io/projected/0b437634-e4b9-4e70-b501-2aee9670efea-kube-api-access-ch548\") pod \"telemetry-operator-controller-manager-76cc84c6bb-zw57g\" (UID: \"0b437634-e4b9-4e70-b501-2aee9670efea\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.645435 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.678755 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.688775 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzmhs\" (UniqueName: \"kubernetes.io/projected/b68334f1-9e8d-48c1-90d0-f650a6c952ac-kube-api-access-wzmhs\") pod \"watcher-operator-controller-manager-769dc69bc-mf6w7\" (UID: \"b68334f1-9e8d-48c1-90d0-f650a6c952ac\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.688825 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dd8j\" (UniqueName: \"kubernetes.io/projected/eec4a71f-5a1e-493f-b820-523ac0712a46-kube-api-access-2dd8j\") pod \"test-operator-controller-manager-5854674fcc-z6qpg\" (UID: \"eec4a71f-5a1e-493f-b820-523ac0712a46\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.761020 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.789593 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7llk7\" (UniqueName: \"kubernetes.io/projected/f2c5de87-7240-4169-8acb-d34261d2479f-kube-api-access-7llk7\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.789661 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.789731 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt9xt\" (UniqueName: \"kubernetes.io/projected/65277cff-497a-4135-988c-94d7456a5dfc-kube-api-access-pt9xt\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qnndl\" (UID: \"65277cff-497a-4135-988c-94d7456a5dfc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.789770 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzmhs\" (UniqueName: \"kubernetes.io/projected/b68334f1-9e8d-48c1-90d0-f650a6c952ac-kube-api-access-wzmhs\") pod \"watcher-operator-controller-manager-769dc69bc-mf6w7\" (UID: \"b68334f1-9e8d-48c1-90d0-f650a6c952ac\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.789811 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dd8j\" (UniqueName: \"kubernetes.io/projected/eec4a71f-5a1e-493f-b820-523ac0712a46-kube-api-access-2dd8j\") pod \"test-operator-controller-manager-5854674fcc-z6qpg\" (UID: \"eec4a71f-5a1e-493f-b820-523ac0712a46\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.789840 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.789861 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.790337 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.790613 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert podName:38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee nodeName:}" failed. No retries permitted until 2025-12-04 17:56:48.790495955 +0000 UTC m=+1070.745856991 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f59cdsz" (UID: "38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.813627 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dd8j\" (UniqueName: \"kubernetes.io/projected/eec4a71f-5a1e-493f-b820-523ac0712a46-kube-api-access-2dd8j\") pod \"test-operator-controller-manager-5854674fcc-z6qpg\" (UID: \"eec4a71f-5a1e-493f-b820-523ac0712a46\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.813604 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzmhs\" (UniqueName: \"kubernetes.io/projected/b68334f1-9e8d-48c1-90d0-f650a6c952ac-kube-api-access-wzmhs\") pod \"watcher-operator-controller-manager-769dc69bc-mf6w7\" (UID: \"b68334f1-9e8d-48c1-90d0-f650a6c952ac\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.836899 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.840000 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.853871 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.859554 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps"] Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.891146 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.891186 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.891211 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7llk7\" (UniqueName: \"kubernetes.io/projected/f2c5de87-7240-4169-8acb-d34261d2479f-kube-api-access-7llk7\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.891297 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt9xt\" (UniqueName: \"kubernetes.io/projected/65277cff-497a-4135-988c-94d7456a5dfc-kube-api-access-pt9xt\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qnndl\" (UID: \"65277cff-497a-4135-988c-94d7456a5dfc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.891703 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.892226 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.892306 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:56:48.392288972 +0000 UTC m=+1070.347650018 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "webhook-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: E1204 17:56:47.892349 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:56:48.392326573 +0000 UTC m=+1070.347687699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "metrics-server-cert" not found Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.912281 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt9xt\" (UniqueName: \"kubernetes.io/projected/65277cff-497a-4135-988c-94d7456a5dfc-kube-api-access-pt9xt\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qnndl\" (UID: \"65277cff-497a-4135-988c-94d7456a5dfc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.914502 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7llk7\" (UniqueName: \"kubernetes.io/projected/f2c5de87-7240-4169-8acb-d34261d2479f-kube-api-access-7llk7\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:47 crc kubenswrapper[4733]: W1204 17:56:47.944635 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e110407_2b5b_4f7a_b815_77203c99b56b.slice/crio-a6ab0a8597c027efbffb73aed8c64db501a14fb987a1b1e5963092133ce5877f WatchSource:0}: Error finding container a6ab0a8597c027efbffb73aed8c64db501a14fb987a1b1e5963092133ce5877f: Status 404 returned error can't find the container with id a6ab0a8597c027efbffb73aed8c64db501a14fb987a1b1e5963092133ce5877f Dec 04 17:56:47 crc kubenswrapper[4733]: I1204 17:56:47.968008 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.000476 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.103236 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz"] Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.113616 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp"] Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.129745 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod056b1e59_a544_4274_ae2e_b3d35ea1ab06.slice/crio-7701e9bcecfc810e0097d23607431484ba53f6f837aedfc3e6a10f3fc3894bf1 WatchSource:0}: Error finding container 7701e9bcecfc810e0097d23607431484ba53f6f837aedfc3e6a10f3fc3894bf1: Status 404 returned error can't find the container with id 7701e9bcecfc810e0097d23607431484ba53f6f837aedfc3e6a10f3fc3894bf1 Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.134037 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50dfe63c_b3b2_46c5_ad8d_91c64e3ef698.slice/crio-d603c6f5485f7b85f70e7e303428a0f2db995118c7ec7f19902128f6a61fe53e WatchSource:0}: Error finding container d603c6f5485f7b85f70e7e303428a0f2db995118c7ec7f19902128f6a61fe53e: Status 404 returned error can't find the container with id d603c6f5485f7b85f70e7e303428a0f2db995118c7ec7f19902128f6a61fe53e Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.241768 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5"] Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.245152 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod143e7c55_1ecb_4efa_8992_332718ae5b61.slice/crio-88b5b1b3c484931d2083620c4420ec1148be99980519ccee85d9431c8d843f76 WatchSource:0}: Error finding container 88b5b1b3c484931d2083620c4420ec1148be99980519ccee85d9431c8d843f76: Status 404 returned error can't find the container with id 88b5b1b3c484931d2083620c4420ec1148be99980519ccee85d9431c8d843f76 Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.282703 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2"] Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.285963 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f29393d_5068_4870_86ab_237c52f6d1f9.slice/crio-33cc4091a72d85089870c62098d66fec6e9973f90865d4e2f21d1c01c9521406 WatchSource:0}: Error finding container 33cc4091a72d85089870c62098d66fec6e9973f90865d4e2f21d1c01c9521406: Status 404 returned error can't find the container with id 33cc4091a72d85089870c62098d66fec6e9973f90865d4e2f21d1c01c9521406 Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.289837 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h"] Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.306982 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z"] Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.309465 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2191192e_388c_4535_be27_5b4b76aea175.slice/crio-3eb741669ec23ad824f171df5633f8923de5b21eea58270c9c10b3c4fcedb547 WatchSource:0}: Error finding container 3eb741669ec23ad824f171df5633f8923de5b21eea58270c9c10b3c4fcedb547: Status 404 returned error can't find the container with id 3eb741669ec23ad824f171df5633f8923de5b21eea58270c9c10b3c4fcedb547 Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.401706 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.401745 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.403038 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.403128 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:56:49.403109904 +0000 UTC m=+1071.358470950 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "webhook-server-cert" not found Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.403491 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.403527 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:56:49.403519684 +0000 UTC m=+1071.358880730 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "metrics-server-cert" not found Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.440376 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0217980_ef7c_4b3a_b5cf_9887545b7a13.slice/crio-2fd223c1581082d1b983c12bd6da3ec29c6eac24051d005382ab78388207c444 WatchSource:0}: Error finding container 2fd223c1581082d1b983c12bd6da3ec29c6eac24051d005382ab78388207c444: Status 404 returned error can't find the container with id 2fd223c1581082d1b983c12bd6da3ec29c6eac24051d005382ab78388207c444 Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.446262 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh"] Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.456819 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc"] Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.469811 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9tvz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-k5gsb_openstack-operators(889320d7-9fc9-43b9-be8b-d5d8ceeb966c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.472331 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9tvz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-k5gsb_openstack-operators(889320d7-9fc9-43b9-be8b-d5d8ceeb966c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.474117 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" podUID="889320d7-9fc9-43b9-be8b-d5d8ceeb966c" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.484837 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48"] Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.496644 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb"] Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.502718 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.502855 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.502899 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert podName:504dcc38-54b7-430f-8c32-e11e5bbbf5dc nodeName:}" failed. No retries permitted until 2025-12-04 17:56:50.502885976 +0000 UTC m=+1072.458247022 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert") pod "infra-operator-controller-manager-57548d458d-8skx7" (UID: "504dcc38-54b7-430f-8c32-e11e5bbbf5dc") : secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.573034 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g"] Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.589280 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm"] Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.589402 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b437634_e4b9_4e70_b501_2aee9670efea.slice/crio-0d4f97276a5a5389d79ba637c00333d087e401334cd7d55379caecb53e4296c6 WatchSource:0}: Error finding container 0d4f97276a5a5389d79ba637c00333d087e401334cd7d55379caecb53e4296c6: Status 404 returned error can't find the container with id 0d4f97276a5a5389d79ba637c00333d087e401334cd7d55379caecb53e4296c6 Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.597784 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w"] Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.603607 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddmnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-7rb9w_openstack-operators(89feaf19-9df3-415a-8bbf-57168c3b89ba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.603706 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xhtvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-txjrm_openstack-operators(e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.605405 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8"] Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.606982 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddmnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-7rb9w_openstack-operators(89feaf19-9df3-415a-8bbf-57168c3b89ba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.607064 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xhtvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-txjrm_openstack-operators(e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.607748 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sxln6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-5rln8_openstack-operators(dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.608654 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" podUID="e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.608737 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" podUID="89feaf19-9df3-415a-8bbf-57168c3b89ba" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.609580 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sxln6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-5rln8_openstack-operators(dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.610966 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" podUID="dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.619999 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl"] Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.634806 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65277cff_497a_4135_988c_94d7456a5dfc.slice/crio-ed75d19a1e48a5f1ad30a168882f5386d1ca1519dacad3a2d4a9fc4bf46534ca WatchSource:0}: Error finding container ed75d19a1e48a5f1ad30a168882f5386d1ca1519dacad3a2d4a9fc4bf46534ca: Status 404 returned error can't find the container with id ed75d19a1e48a5f1ad30a168882f5386d1ca1519dacad3a2d4a9fc4bf46534ca Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.637434 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb68334f1_9e8d_48c1_90d0_f650a6c952ac.slice/crio-c3d2b2f1da9f298b7d30be80a1d1c1901d4f365b10c61923c51d25d51a2a928e WatchSource:0}: Error finding container c3d2b2f1da9f298b7d30be80a1d1c1901d4f365b10c61923c51d25d51a2a928e: Status 404 returned error can't find the container with id c3d2b2f1da9f298b7d30be80a1d1c1901d4f365b10c61923c51d25d51a2a928e Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.638282 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pt9xt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qnndl_openstack-operators(65277cff-497a-4135-988c-94d7456a5dfc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.639374 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" podUID="65277cff-497a-4135-988c-94d7456a5dfc" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.639675 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7"] Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.639972 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzmhs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-mf6w7_openstack-operators(b68334f1-9e8d-48c1-90d0-f650a6c952ac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: W1204 17:56:48.642024 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeec4a71f_5a1e_493f_b820_523ac0712a46.slice/crio-50989f2217792227c441502f9d64a886a56b58ef934f7132e0d3819860705b10 WatchSource:0}: Error finding container 50989f2217792227c441502f9d64a886a56b58ef934f7132e0d3819860705b10: Status 404 returned error can't find the container with id 50989f2217792227c441502f9d64a886a56b58ef934f7132e0d3819860705b10 Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.642119 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzmhs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-mf6w7_openstack-operators(b68334f1-9e8d-48c1-90d0-f650a6c952ac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.643286 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" podUID="b68334f1-9e8d-48c1-90d0-f650a6c952ac" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.644649 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2dd8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-z6qpg_openstack-operators(eec4a71f-5a1e-493f-b820-523ac0712a46): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.647326 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2dd8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-z6qpg_openstack-operators(eec4a71f-5a1e-493f-b820-523ac0712a46): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.648514 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" podUID="eec4a71f-5a1e-493f-b820-523ac0712a46" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.657455 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg"] Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.667359 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" event={"ID":"889320d7-9fc9-43b9-be8b-d5d8ceeb966c","Type":"ContainerStarted","Data":"cec33077e9c1f963e44a5b6b7dbe7569157da0b4a12e3fb5dbacfb2cba917442"} Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.669211 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" podUID="889320d7-9fc9-43b9-be8b-d5d8ceeb966c" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.671207 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" event={"ID":"e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5","Type":"ContainerStarted","Data":"e2b6740d5c27bbbe7e031a5f2a64ef54fdec10beddfd2a377a54cb6d6e1fd8c1"} Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.672686 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" podUID="e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.673314 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" event={"ID":"65277cff-497a-4135-988c-94d7456a5dfc","Type":"ContainerStarted","Data":"ed75d19a1e48a5f1ad30a168882f5386d1ca1519dacad3a2d4a9fc4bf46534ca"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.674707 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" event={"ID":"fbad1f3f-55ee-4477-a8a4-c146ee09ba51","Type":"ContainerStarted","Data":"8cbecc58c1601298931ec8814afef8f4bc34cd6cb7ce050eac19495edab8c526"} Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.675027 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" podUID="65277cff-497a-4135-988c-94d7456a5dfc" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.676896 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" event={"ID":"2191192e-388c-4535-be27-5b4b76aea175","Type":"ContainerStarted","Data":"3eb741669ec23ad824f171df5633f8923de5b21eea58270c9c10b3c4fcedb547"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.677937 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" event={"ID":"da343538-3e4a-49a0-949d-2622d62ed6a2","Type":"ContainerStarted","Data":"495e47060568cf5fa3b01c822ac203865bc9be5da14dde51a0274685e90827d5"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.690767 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" event={"ID":"dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8","Type":"ContainerStarted","Data":"7c2a4e7eb5e243243a156ed1f3bc575c4bbf2cbe5489f9c0d29c943da6201816"} Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.693966 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" podUID="dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.695097 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" event={"ID":"6c895222-dbc2-4490-972c-b7e9dad9e839","Type":"ContainerStarted","Data":"fb6a1da1f2965d3c79ac13f0ec5d2f18e6f95b9c013ceb1e00b958e28f789342"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.696181 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" event={"ID":"d0217980-ef7c-4b3a-b5cf-9887545b7a13","Type":"ContainerStarted","Data":"2fd223c1581082d1b983c12bd6da3ec29c6eac24051d005382ab78388207c444"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.705871 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" event={"ID":"50dfe63c-b3b2-46c5-ad8d-91c64e3ef698","Type":"ContainerStarted","Data":"d603c6f5485f7b85f70e7e303428a0f2db995118c7ec7f19902128f6a61fe53e"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.713717 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" event={"ID":"500a0dd4-904a-449e-956a-013bed8c66ca","Type":"ContainerStarted","Data":"955fc8bd35b33cc03c355c7732d9e1bec05957514179bf5439413964e21d55fb"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.715406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" event={"ID":"9e110407-2b5b-4f7a-b815-77203c99b56b","Type":"ContainerStarted","Data":"a6ab0a8597c027efbffb73aed8c64db501a14fb987a1b1e5963092133ce5877f"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.722308 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" event={"ID":"f4fddbf6-0bc5-4b36-9a8a-7073e875c540","Type":"ContainerStarted","Data":"e97a82508ae1ef7003555966596fce9960e62c7f8888115def3bd9ed28b86c70"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.733738 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" event={"ID":"7f29393d-5068-4870-86ab-237c52f6d1f9","Type":"ContainerStarted","Data":"33cc4091a72d85089870c62098d66fec6e9973f90865d4e2f21d1c01c9521406"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.736485 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" event={"ID":"89feaf19-9df3-415a-8bbf-57168c3b89ba","Type":"ContainerStarted","Data":"3d7aa27b263c9b33c00d9215b1cc21a4cf4a9840a82cbe47bc4f69cf71da4d5e"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.740106 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" event={"ID":"143e7c55-1ecb-4efa-8992-332718ae5b61","Type":"ContainerStarted","Data":"88b5b1b3c484931d2083620c4420ec1148be99980519ccee85d9431c8d843f76"} Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.749848 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" podUID="89feaf19-9df3-415a-8bbf-57168c3b89ba" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.752831 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" event={"ID":"0b437634-e4b9-4e70-b501-2aee9670efea","Type":"ContainerStarted","Data":"0d4f97276a5a5389d79ba637c00333d087e401334cd7d55379caecb53e4296c6"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.756496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" event={"ID":"eec4a71f-5a1e-493f-b820-523ac0712a46","Type":"ContainerStarted","Data":"50989f2217792227c441502f9d64a886a56b58ef934f7132e0d3819860705b10"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.760075 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" event={"ID":"b68334f1-9e8d-48c1-90d0-f650a6c952ac","Type":"ContainerStarted","Data":"c3d2b2f1da9f298b7d30be80a1d1c1901d4f365b10c61923c51d25d51a2a928e"} Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.798571 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" podUID="b68334f1-9e8d-48c1-90d0-f650a6c952ac" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.799868 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" podUID="eec4a71f-5a1e-493f-b820-523ac0712a46" Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.802252 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" event={"ID":"056b1e59-a544-4274-ae2e-b3d35ea1ab06","Type":"ContainerStarted","Data":"7701e9bcecfc810e0097d23607431484ba53f6f837aedfc3e6a10f3fc3894bf1"} Dec 04 17:56:48 crc kubenswrapper[4733]: I1204 17:56:48.807924 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.808874 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:48 crc kubenswrapper[4733]: E1204 17:56:48.808913 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert podName:38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee nodeName:}" failed. No retries permitted until 2025-12-04 17:56:50.808900382 +0000 UTC m=+1072.764261428 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f59cdsz" (UID: "38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:49 crc kubenswrapper[4733]: I1204 17:56:49.423161 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:49 crc kubenswrapper[4733]: I1204 17:56:49.423222 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.423936 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.423994 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:56:51.423980227 +0000 UTC m=+1073.379341273 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "webhook-server-cert" not found Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.423986 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.424091 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:56:51.424063009 +0000 UTC m=+1073.379424105 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "metrics-server-cert" not found Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.815833 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" podUID="65277cff-497a-4135-988c-94d7456a5dfc" Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.816495 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" podUID="889320d7-9fc9-43b9-be8b-d5d8ceeb966c" Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.816563 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" podUID="89feaf19-9df3-415a-8bbf-57168c3b89ba" Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.816923 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" podUID="e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5" Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.816970 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" podUID="b68334f1-9e8d-48c1-90d0-f650a6c952ac" Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.820053 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" podUID="eec4a71f-5a1e-493f-b820-523ac0712a46" Dec 04 17:56:49 crc kubenswrapper[4733]: E1204 17:56:49.820314 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" podUID="dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8" Dec 04 17:56:50 crc kubenswrapper[4733]: I1204 17:56:50.553937 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:50 crc kubenswrapper[4733]: E1204 17:56:50.554081 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:50 crc kubenswrapper[4733]: E1204 17:56:50.554128 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert podName:504dcc38-54b7-430f-8c32-e11e5bbbf5dc nodeName:}" failed. No retries permitted until 2025-12-04 17:56:54.554113597 +0000 UTC m=+1076.509474643 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert") pod "infra-operator-controller-manager-57548d458d-8skx7" (UID: "504dcc38-54b7-430f-8c32-e11e5bbbf5dc") : secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:50 crc kubenswrapper[4733]: I1204 17:56:50.861145 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:50 crc kubenswrapper[4733]: E1204 17:56:50.861358 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:50 crc kubenswrapper[4733]: E1204 17:56:50.861411 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert podName:38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee nodeName:}" failed. No retries permitted until 2025-12-04 17:56:54.861395618 +0000 UTC m=+1076.816756664 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f59cdsz" (UID: "38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:51 crc kubenswrapper[4733]: I1204 17:56:51.482723 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:51 crc kubenswrapper[4733]: I1204 17:56:51.483179 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:51 crc kubenswrapper[4733]: E1204 17:56:51.483544 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 17:56:51 crc kubenswrapper[4733]: E1204 17:56:51.483632 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:56:55.483618484 +0000 UTC m=+1077.438979530 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "webhook-server-cert" not found Dec 04 17:56:51 crc kubenswrapper[4733]: E1204 17:56:51.484161 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 17:56:51 crc kubenswrapper[4733]: E1204 17:56:51.484215 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:56:55.484184899 +0000 UTC m=+1077.439545955 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "metrics-server-cert" not found Dec 04 17:56:54 crc kubenswrapper[4733]: I1204 17:56:54.638987 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:56:54 crc kubenswrapper[4733]: E1204 17:56:54.639288 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:54 crc kubenswrapper[4733]: E1204 17:56:54.639533 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert podName:504dcc38-54b7-430f-8c32-e11e5bbbf5dc nodeName:}" failed. No retries permitted until 2025-12-04 17:57:02.639494779 +0000 UTC m=+1084.594855825 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert") pod "infra-operator-controller-manager-57548d458d-8skx7" (UID: "504dcc38-54b7-430f-8c32-e11e5bbbf5dc") : secret "infra-operator-webhook-server-cert" not found Dec 04 17:56:54 crc kubenswrapper[4733]: I1204 17:56:54.942893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:56:54 crc kubenswrapper[4733]: E1204 17:56:54.943040 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:54 crc kubenswrapper[4733]: E1204 17:56:54.943112 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert podName:38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee nodeName:}" failed. No retries permitted until 2025-12-04 17:57:02.943093911 +0000 UTC m=+1084.898454957 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f59cdsz" (UID: "38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 17:56:55 crc kubenswrapper[4733]: I1204 17:56:55.549815 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:55 crc kubenswrapper[4733]: I1204 17:56:55.549862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:56:55 crc kubenswrapper[4733]: E1204 17:56:55.550017 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 17:56:55 crc kubenswrapper[4733]: E1204 17:56:55.550046 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 17:56:55 crc kubenswrapper[4733]: E1204 17:56:55.550070 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:57:03.550056917 +0000 UTC m=+1085.505417963 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "webhook-server-cert" not found Dec 04 17:56:55 crc kubenswrapper[4733]: E1204 17:56:55.550178 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs podName:f2c5de87-7240-4169-8acb-d34261d2479f nodeName:}" failed. No retries permitted until 2025-12-04 17:57:03.550150789 +0000 UTC m=+1085.505511875 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-gblbb" (UID: "f2c5de87-7240-4169-8acb-d34261d2479f") : secret "metrics-server-cert" not found Dec 04 17:57:01 crc kubenswrapper[4733]: I1204 17:57:01.337129 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 17:57:01 crc kubenswrapper[4733]: E1204 17:57:01.742672 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 04 17:57:01 crc kubenswrapper[4733]: E1204 17:57:01.742866 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jnctp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-7qmwh_openstack-operators(6c895222-dbc2-4490-972c-b7e9dad9e839): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:57:02 crc kubenswrapper[4733]: I1204 17:57:02.650610 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:57:02 crc kubenswrapper[4733]: I1204 17:57:02.659335 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/504dcc38-54b7-430f-8c32-e11e5bbbf5dc-cert\") pod \"infra-operator-controller-manager-57548d458d-8skx7\" (UID: \"504dcc38-54b7-430f-8c32-e11e5bbbf5dc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:57:02 crc kubenswrapper[4733]: I1204 17:57:02.763286 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-m695k" Dec 04 17:57:02 crc kubenswrapper[4733]: I1204 17:57:02.772270 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:57:02 crc kubenswrapper[4733]: I1204 17:57:02.953864 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:57:02 crc kubenswrapper[4733]: I1204 17:57:02.957386 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f59cdsz\" (UID: \"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:57:03 crc kubenswrapper[4733]: I1204 17:57:03.064220 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-pfn2q" Dec 04 17:57:03 crc kubenswrapper[4733]: I1204 17:57:03.072644 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:57:03 crc kubenswrapper[4733]: I1204 17:57:03.563214 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:57:03 crc kubenswrapper[4733]: I1204 17:57:03.563319 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:57:03 crc kubenswrapper[4733]: I1204 17:57:03.569646 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:57:03 crc kubenswrapper[4733]: I1204 17:57:03.570431 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2c5de87-7240-4169-8acb-d34261d2479f-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-gblbb\" (UID: \"f2c5de87-7240-4169-8acb-d34261d2479f\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:57:03 crc kubenswrapper[4733]: I1204 17:57:03.588257 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-4cxqg" Dec 04 17:57:03 crc kubenswrapper[4733]: I1204 17:57:03.596699 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:57:08 crc kubenswrapper[4733]: E1204 17:57:08.510879 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 04 17:57:08 crc kubenswrapper[4733]: E1204 17:57:08.511444 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7g5gs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-bjzvv_openstack-operators(fbad1f3f-55ee-4477-a8a4-c146ee09ba51): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:57:08 crc kubenswrapper[4733]: E1204 17:57:08.534751 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 04 17:57:08 crc kubenswrapper[4733]: E1204 17:57:08.535092 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2djhv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-v8d4h_openstack-operators(da343538-3e4a-49a0-949d-2622d62ed6a2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:57:09 crc kubenswrapper[4733]: E1204 17:57:09.607628 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 04 17:57:09 crc kubenswrapper[4733]: E1204 17:57:09.608073 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w6mhz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-25jzc_openstack-operators(f4fddbf6-0bc5-4b36-9a8a-7073e875c540): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:57:10 crc kubenswrapper[4733]: E1204 17:57:10.223546 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 04 17:57:10 crc kubenswrapper[4733]: E1204 17:57:10.223733 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tzkp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-qpssp_openstack-operators(50dfe63c-b3b2-46c5-ad8d-91c64e3ef698): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:57:14 crc kubenswrapper[4733]: I1204 17:57:14.243668 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb"] Dec 04 17:57:15 crc kubenswrapper[4733]: I1204 17:57:15.305724 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz"] Dec 04 17:57:15 crc kubenswrapper[4733]: I1204 17:57:15.616063 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-8skx7"] Dec 04 17:57:15 crc kubenswrapper[4733]: W1204 17:57:15.937633 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38cad43f_c4ed_4dcf_8de2_6f86aa9ba4ee.slice/crio-806b0f40fd7dc28765205c3224587c3630d24651fa786cc781cd5d6d2801f034 WatchSource:0}: Error finding container 806b0f40fd7dc28765205c3224587c3630d24651fa786cc781cd5d6d2801f034: Status 404 returned error can't find the container with id 806b0f40fd7dc28765205c3224587c3630d24651fa786cc781cd5d6d2801f034 Dec 04 17:57:15 crc kubenswrapper[4733]: W1204 17:57:15.941870 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod504dcc38_54b7_430f_8c32_e11e5bbbf5dc.slice/crio-10bc3cbaa3b632fe85c770b066c86f901fa405edb47f44ce9330393e5eeb788f WatchSource:0}: Error finding container 10bc3cbaa3b632fe85c770b066c86f901fa405edb47f44ce9330393e5eeb788f: Status 404 returned error can't find the container with id 10bc3cbaa3b632fe85c770b066c86f901fa405edb47f44ce9330393e5eeb788f Dec 04 17:57:16 crc kubenswrapper[4733]: I1204 17:57:16.025337 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" event={"ID":"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee","Type":"ContainerStarted","Data":"806b0f40fd7dc28765205c3224587c3630d24651fa786cc781cd5d6d2801f034"} Dec 04 17:57:16 crc kubenswrapper[4733]: I1204 17:57:16.027209 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" event={"ID":"f2c5de87-7240-4169-8acb-d34261d2479f","Type":"ContainerStarted","Data":"d7670984f019599d1a5bb535609dc86726f1d8abac3867f67537a3adca514216"} Dec 04 17:57:16 crc kubenswrapper[4733]: I1204 17:57:16.028885 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" event={"ID":"504dcc38-54b7-430f-8c32-e11e5bbbf5dc","Type":"ContainerStarted","Data":"10bc3cbaa3b632fe85c770b066c86f901fa405edb47f44ce9330393e5eeb788f"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.098075 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" event={"ID":"056b1e59-a544-4274-ae2e-b3d35ea1ab06","Type":"ContainerStarted","Data":"e44c983aee2194b35ca87e6e66ba900dea5cb88149f4e23ac554d692a5aee8b0"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.122075 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" event={"ID":"d0217980-ef7c-4b3a-b5cf-9887545b7a13","Type":"ContainerStarted","Data":"1ef1873a2370425225c7ccd80cae2c453dfae1d0983b222b076cf85d5871cf7f"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.167319 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" event={"ID":"0b437634-e4b9-4e70-b501-2aee9670efea","Type":"ContainerStarted","Data":"5959eb7660cf82f515ade37508033f0c5b81bcc0dbf7eff582d6b6fa3a56d91a"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.172510 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" event={"ID":"f2c5de87-7240-4169-8acb-d34261d2479f","Type":"ContainerStarted","Data":"5c4814655f3be5cf764b27435074aa890f5faecbfd38c6a44f837aa61eabea4f"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.173346 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.195553 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" event={"ID":"eec4a71f-5a1e-493f-b820-523ac0712a46","Type":"ContainerStarted","Data":"20b32631be63177a1a885646ee8276271163992ed14048474a710e2c3eccdd9a"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.219980 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" event={"ID":"2191192e-388c-4535-be27-5b4b76aea175","Type":"ContainerStarted","Data":"ffb9a1f003e39b033edb509aa2e534872cdca53c8187ac64322011dad9451576"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.233580 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" event={"ID":"500a0dd4-904a-449e-956a-013bed8c66ca","Type":"ContainerStarted","Data":"f9ff264daeda5e3db57eacfda7f3b3cf581cbe6552061452c3181117897441f4"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.242452 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" event={"ID":"7f29393d-5068-4870-86ab-237c52f6d1f9","Type":"ContainerStarted","Data":"a835fca29686462cb682330b6b95689a540adbffbacf047dd00e8a71dd8399a4"} Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.247358 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" podStartSLOduration=30.247344135 podStartE2EDuration="30.247344135s" podCreationTimestamp="2025-12-04 17:56:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:57:17.244138728 +0000 UTC m=+1099.199499774" watchObservedRunningTime="2025-12-04 17:57:17.247344135 +0000 UTC m=+1099.202705181" Dec 04 17:57:17 crc kubenswrapper[4733]: I1204 17:57:17.268470 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" event={"ID":"143e7c55-1ecb-4efa-8992-332718ae5b61","Type":"ContainerStarted","Data":"c767c0b0a1484cd357a442dfb51fbf2307da2d58d812a5fc9c673bfc7fceb067"} Dec 04 17:57:18 crc kubenswrapper[4733]: I1204 17:57:18.297374 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" event={"ID":"9e110407-2b5b-4f7a-b815-77203c99b56b","Type":"ContainerStarted","Data":"bf44e5945b640665822bf57cb684648cba8985fbe1592295a23d518e375235a8"} Dec 04 17:57:18 crc kubenswrapper[4733]: I1204 17:57:18.298594 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" event={"ID":"e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5","Type":"ContainerStarted","Data":"5911317e5968cccadeed3ae211de2876b5bc332a4e6dc249b95dd32801ad0b44"} Dec 04 17:57:18 crc kubenswrapper[4733]: I1204 17:57:18.299705 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" event={"ID":"889320d7-9fc9-43b9-be8b-d5d8ceeb966c","Type":"ContainerStarted","Data":"dc2d0ee07e7502a687cbd65736f1471c5858f14900fe1b24aa8d5b2b79512350"} Dec 04 17:57:20 crc kubenswrapper[4733]: E1204 17:57:20.202754 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" podUID="50dfe63c-b3b2-46c5-ad8d-91c64e3ef698" Dec 04 17:57:20 crc kubenswrapper[4733]: E1204 17:57:20.237451 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" podUID="da343538-3e4a-49a0-949d-2622d62ed6a2" Dec 04 17:57:20 crc kubenswrapper[4733]: I1204 17:57:20.351225 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" event={"ID":"89feaf19-9df3-415a-8bbf-57168c3b89ba","Type":"ContainerStarted","Data":"026fd77955d6eb41e06e734bbdca1e20415e7152d2841a597dbc35cacb2f2be3"} Dec 04 17:57:20 crc kubenswrapper[4733]: I1204 17:57:20.351644 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" event={"ID":"da343538-3e4a-49a0-949d-2622d62ed6a2","Type":"ContainerStarted","Data":"d297a8867aba5f55bf8acab79ced00eb954a4b65689dfccf1378161861716456"} Dec 04 17:57:20 crc kubenswrapper[4733]: I1204 17:57:20.359281 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" event={"ID":"65277cff-497a-4135-988c-94d7456a5dfc","Type":"ContainerStarted","Data":"5bd61c0100e6f60b53c45fc858b4755b183421423d1215bd65e581707fa6a806"} Dec 04 17:57:20 crc kubenswrapper[4733]: I1204 17:57:20.384186 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" event={"ID":"dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8","Type":"ContainerStarted","Data":"9cadb794ba9c6d70039d5692cc2dd42a656fc139c18686d58045776828bf63c9"} Dec 04 17:57:20 crc kubenswrapper[4733]: I1204 17:57:20.409026 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" event={"ID":"50dfe63c-b3b2-46c5-ad8d-91c64e3ef698","Type":"ContainerStarted","Data":"524120ab574c717ca178eac3b813cd115a5f004bf2819d0ed2125f303e48763e"} Dec 04 17:57:20 crc kubenswrapper[4733]: I1204 17:57:20.427655 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qnndl" podStartSLOduration=5.546551007 podStartE2EDuration="33.427636846s" podCreationTimestamp="2025-12-04 17:56:47 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.638146562 +0000 UTC m=+1070.593507608" lastFinishedPulling="2025-12-04 17:57:16.519232401 +0000 UTC m=+1098.474593447" observedRunningTime="2025-12-04 17:57:20.423264199 +0000 UTC m=+1102.378625245" watchObservedRunningTime="2025-12-04 17:57:20.427636846 +0000 UTC m=+1102.382997892" Dec 04 17:57:20 crc kubenswrapper[4733]: I1204 17:57:20.437888 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" event={"ID":"b68334f1-9e8d-48c1-90d0-f650a6c952ac","Type":"ContainerStarted","Data":"00ac07dfd7c7359c20335d22627a724a517d37beaefee2cc03c3ebb114fbe813"} Dec 04 17:57:20 crc kubenswrapper[4733]: E1204 17:57:20.510280 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" podUID="fbad1f3f-55ee-4477-a8a4-c146ee09ba51" Dec 04 17:57:20 crc kubenswrapper[4733]: E1204 17:57:20.525091 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" podUID="6c895222-dbc2-4490-972c-b7e9dad9e839" Dec 04 17:57:20 crc kubenswrapper[4733]: E1204 17:57:20.921030 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" podUID="f4fddbf6-0bc5-4b36-9a8a-7073e875c540" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.467301 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" event={"ID":"500a0dd4-904a-449e-956a-013bed8c66ca","Type":"ContainerStarted","Data":"a26072a2303f54b20ff0719810ef2b82c500ae3d6c2cb11b648ac686402a5607"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.469211 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.474200 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.489643 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" event={"ID":"7f29393d-5068-4870-86ab-237c52f6d1f9","Type":"ContainerStarted","Data":"81a0a848a1206474ebc4ec2b1c620e461e91e04cb46cc7ba95a1340ef18edeff"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.490395 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.492211 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.504453 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-8t88q" podStartSLOduration=3.261841434 podStartE2EDuration="35.504433177s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:47.87995253 +0000 UTC m=+1069.835313576" lastFinishedPulling="2025-12-04 17:57:20.122544273 +0000 UTC m=+1102.077905319" observedRunningTime="2025-12-04 17:57:21.494056412 +0000 UTC m=+1103.449417458" watchObservedRunningTime="2025-12-04 17:57:21.504433177 +0000 UTC m=+1103.459794223" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.508114 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" event={"ID":"e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5","Type":"ContainerStarted","Data":"5e0f8935570347a181d50e36f0d42439da86cd17da5a017db7f4b37b4b15c85e"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.508258 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.516196 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" event={"ID":"dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8","Type":"ContainerStarted","Data":"44d899d47dd792bf37b40c0e0ac958fb6c82c53541e5dca1230fec24696bc66d"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.516887 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.517979 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" event={"ID":"eec4a71f-5a1e-493f-b820-523ac0712a46","Type":"ContainerStarted","Data":"c1fa4183c28c301a16b8eb692968430e74d6a924946153361b22c18635ccf465"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.518528 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.521638 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.541946 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8r2s2" podStartSLOduration=3.896486188 podStartE2EDuration="35.541930474s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.289475759 +0000 UTC m=+1070.244836805" lastFinishedPulling="2025-12-04 17:57:19.934920045 +0000 UTC m=+1101.890281091" observedRunningTime="2025-12-04 17:57:21.541851562 +0000 UTC m=+1103.497212618" watchObservedRunningTime="2025-12-04 17:57:21.541930474 +0000 UTC m=+1103.497291520" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.563935 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" event={"ID":"056b1e59-a544-4274-ae2e-b3d35ea1ab06","Type":"ContainerStarted","Data":"5ace76398e98c2bad81f436d0e78acd3de4f2e01acb86ff66727957b07bf47bc"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.564780 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.566946 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.571057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" event={"ID":"143e7c55-1ecb-4efa-8992-332718ae5b61","Type":"ContainerStarted","Data":"ac120a01e4d2365430d7ec283897c6d844e40c058d3a007fc0478219857d57be"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.572000 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.578749 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.580496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" event={"ID":"0b437634-e4b9-4e70-b501-2aee9670efea","Type":"ContainerStarted","Data":"0c4465518478c0aa6db6dfae209cb3820312e476f3747cc2028bfb861709cf30"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.580591 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6qpg" podStartSLOduration=3.203454634 podStartE2EDuration="34.580568181s" podCreationTimestamp="2025-12-04 17:56:47 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.644555394 +0000 UTC m=+1070.599916440" lastFinishedPulling="2025-12-04 17:57:20.021668911 +0000 UTC m=+1101.977029987" observedRunningTime="2025-12-04 17:57:21.57034203 +0000 UTC m=+1103.525703076" watchObservedRunningTime="2025-12-04 17:57:21.580568181 +0000 UTC m=+1103.535929227" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.581254 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.592754 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.607548 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" podStartSLOduration=4.065362601 podStartE2EDuration="35.607534458s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.603653915 +0000 UTC m=+1070.559014961" lastFinishedPulling="2025-12-04 17:57:20.145825772 +0000 UTC m=+1102.101186818" observedRunningTime="2025-12-04 17:57:21.603993354 +0000 UTC m=+1103.559354410" watchObservedRunningTime="2025-12-04 17:57:21.607534458 +0000 UTC m=+1103.562895504" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.612601 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" event={"ID":"6c895222-dbc2-4490-972c-b7e9dad9e839","Type":"ContainerStarted","Data":"97976e58ac2ec756c03d152a410ec387d14cbe2980320b0d4238581f8792a82a"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.633325 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" event={"ID":"2191192e-388c-4535-be27-5b4b76aea175","Type":"ContainerStarted","Data":"172a6b295bf41d35d505f16fc49da50ebdd2aeee45fb3be50866c93d4cc06560"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.634203 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.643097 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.657074 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" event={"ID":"889320d7-9fc9-43b9-be8b-d5d8ceeb966c","Type":"ContainerStarted","Data":"a4fdb3d1d8f3a700cf8b03aef36748518ff623951a50dd8f47e806038e312ade"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.657222 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.681493 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" podStartSLOduration=8.341665905 podStartE2EDuration="35.681470824s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.607657312 +0000 UTC m=+1070.563018358" lastFinishedPulling="2025-12-04 17:57:15.947462231 +0000 UTC m=+1097.902823277" observedRunningTime="2025-12-04 17:57:21.635992795 +0000 UTC m=+1103.591353841" watchObservedRunningTime="2025-12-04 17:57:21.681470824 +0000 UTC m=+1103.636831870" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.681749 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vqnt5" podStartSLOduration=3.994954345 podStartE2EDuration="35.681732372s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.247923692 +0000 UTC m=+1070.203284738" lastFinishedPulling="2025-12-04 17:57:19.934701709 +0000 UTC m=+1101.890062765" observedRunningTime="2025-12-04 17:57:21.660410614 +0000 UTC m=+1103.615771670" watchObservedRunningTime="2025-12-04 17:57:21.681732372 +0000 UTC m=+1103.637093418" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.700770 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" event={"ID":"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee","Type":"ContainerStarted","Data":"b3041afc8dfa6f12ec44cb892794d00e3f07f1d054240d6f9a33317f6ede309f"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.701032 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" event={"ID":"38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee","Type":"ContainerStarted","Data":"786ec0341e1d16d78754ba6419ad267b1e5eb2c38f2ead42b92cc9a158bd1479"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.701618 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.717082 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" event={"ID":"f4fddbf6-0bc5-4b36-9a8a-7073e875c540","Type":"ContainerStarted","Data":"59b1667a4cdd5b2b189ab37f933acb605c8e9cb10240034510a99d3e4f5b6adf"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.725665 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cs6cz" podStartSLOduration=3.836434946 podStartE2EDuration="35.725648349s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.132440528 +0000 UTC m=+1070.087801564" lastFinishedPulling="2025-12-04 17:57:20.021653911 +0000 UTC m=+1101.977014967" observedRunningTime="2025-12-04 17:57:21.725091164 +0000 UTC m=+1103.680452200" watchObservedRunningTime="2025-12-04 17:57:21.725648349 +0000 UTC m=+1103.681009395" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.727779 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2pm4z" podStartSLOduration=4.136278192 podStartE2EDuration="35.727773295s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.315132059 +0000 UTC m=+1070.270493105" lastFinishedPulling="2025-12-04 17:57:19.906627162 +0000 UTC m=+1101.861988208" observedRunningTime="2025-12-04 17:57:21.693373311 +0000 UTC m=+1103.648734357" watchObservedRunningTime="2025-12-04 17:57:21.727773295 +0000 UTC m=+1103.683134341" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.739098 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" event={"ID":"d0217980-ef7c-4b3a-b5cf-9887545b7a13","Type":"ContainerStarted","Data":"f69caaad6de880a6bbea2ba993fd8e12d2ec9dcbb9898b4f1959c58a1b7d530c"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.740101 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.744964 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.751201 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" event={"ID":"89feaf19-9df3-415a-8bbf-57168c3b89ba","Type":"ContainerStarted","Data":"b3839c124634f9828dc5f25e684deafd36deac7a7567d1153a73d028aff73faa"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.752011 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.754847 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-zw57g" podStartSLOduration=3.346913186 podStartE2EDuration="34.754838086s" podCreationTimestamp="2025-12-04 17:56:47 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.601442355 +0000 UTC m=+1070.556803401" lastFinishedPulling="2025-12-04 17:57:20.009367255 +0000 UTC m=+1101.964728301" observedRunningTime="2025-12-04 17:57:21.750962582 +0000 UTC m=+1103.706323628" watchObservedRunningTime="2025-12-04 17:57:21.754838086 +0000 UTC m=+1103.710199132" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.761961 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" event={"ID":"9e110407-2b5b-4f7a-b815-77203c99b56b","Type":"ContainerStarted","Data":"d345365154a472a1450858bdd8a74372f0a3317e8157adde62348be964309587"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.762426 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.775967 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" event={"ID":"fbad1f3f-55ee-4477-a8a4-c146ee09ba51","Type":"ContainerStarted","Data":"0c7f7782d910e4cbfed99634455258134edab8ed62b5b5415184e1d68ea8d717"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.793119 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" event={"ID":"b68334f1-9e8d-48c1-90d0-f650a6c952ac","Type":"ContainerStarted","Data":"70b0d7ba34ad40940503e2fcd7b827aa1871b71ba4d8280dc52aced750a765e6"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.793836 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.805922 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" event={"ID":"504dcc38-54b7-430f-8c32-e11e5bbbf5dc","Type":"ContainerStarted","Data":"e8edf1f46f2eac2904062fa07dbe5de55dc899de3f3b964bfd9a3a56ee823f84"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.805963 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" event={"ID":"504dcc38-54b7-430f-8c32-e11e5bbbf5dc","Type":"ContainerStarted","Data":"4993c8d427ee0ef5938de39d82a1c39ac83e3389909fc8769fc5fc5fa5f4b366"} Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.806573 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.855263 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" podStartSLOduration=3.785839856 podStartE2EDuration="35.855242754s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:47.949620383 +0000 UTC m=+1069.904981429" lastFinishedPulling="2025-12-04 17:57:20.019023251 +0000 UTC m=+1101.974384327" observedRunningTime="2025-12-04 17:57:21.838109799 +0000 UTC m=+1103.793470835" watchObservedRunningTime="2025-12-04 17:57:21.855242754 +0000 UTC m=+1103.810603790" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.911853 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" podStartSLOduration=4.258396625 podStartE2EDuration="35.911808469s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.469665963 +0000 UTC m=+1070.425027009" lastFinishedPulling="2025-12-04 17:57:20.123077807 +0000 UTC m=+1102.078438853" observedRunningTime="2025-12-04 17:57:21.898768012 +0000 UTC m=+1103.854129068" watchObservedRunningTime="2025-12-04 17:57:21.911808469 +0000 UTC m=+1103.867169525" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.934594 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" podStartSLOduration=8.590755769 podStartE2EDuration="35.934576845s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.603493111 +0000 UTC m=+1070.558854157" lastFinishedPulling="2025-12-04 17:57:15.947314167 +0000 UTC m=+1097.902675233" observedRunningTime="2025-12-04 17:57:21.92766434 +0000 UTC m=+1103.883025406" watchObservedRunningTime="2025-12-04 17:57:21.934576845 +0000 UTC m=+1103.889937891" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.951963 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-v8s48" podStartSLOduration=4.38532344 podStartE2EDuration="35.951951756s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.443097988 +0000 UTC m=+1070.398459034" lastFinishedPulling="2025-12-04 17:57:20.009726304 +0000 UTC m=+1101.965087350" observedRunningTime="2025-12-04 17:57:21.946916292 +0000 UTC m=+1103.902277338" watchObservedRunningTime="2025-12-04 17:57:21.951951756 +0000 UTC m=+1103.907312802" Dec 04 17:57:21 crc kubenswrapper[4733]: I1204 17:57:21.994592 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" podStartSLOduration=32.132920788 podStartE2EDuration="35.994578029s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:57:15.941161531 +0000 UTC m=+1097.896522587" lastFinishedPulling="2025-12-04 17:57:19.802818782 +0000 UTC m=+1101.758179828" observedRunningTime="2025-12-04 17:57:21.992786272 +0000 UTC m=+1103.948147318" watchObservedRunningTime="2025-12-04 17:57:21.994578029 +0000 UTC m=+1103.949939075" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.025995 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" podStartSLOduration=8.425059119 podStartE2EDuration="35.025980854s" podCreationTimestamp="2025-12-04 17:56:47 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.639895369 +0000 UTC m=+1070.595256415" lastFinishedPulling="2025-12-04 17:57:15.240817104 +0000 UTC m=+1097.196178150" observedRunningTime="2025-12-04 17:57:22.016490912 +0000 UTC m=+1103.971851948" watchObservedRunningTime="2025-12-04 17:57:22.025980854 +0000 UTC m=+1103.981341900" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.825122 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" event={"ID":"da343538-3e4a-49a0-949d-2622d62ed6a2","Type":"ContainerStarted","Data":"f905ec71e49b89ca2a429c55e2f2f27fd3498951c984f38b0d98da7f3a2af2c8"} Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.825212 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.827152 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" event={"ID":"fbad1f3f-55ee-4477-a8a4-c146ee09ba51","Type":"ContainerStarted","Data":"041c915e6465aefc27f5d143a7a515046a5a0ef753f162244bd70896c8b57ba5"} Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.827830 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.829148 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" event={"ID":"6c895222-dbc2-4490-972c-b7e9dad9e839","Type":"ContainerStarted","Data":"d447f1aa1a0047930e5b68914e2543f1fc12e6cb122aa6bcb6c2af3b4e610bc8"} Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.829641 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.831556 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" event={"ID":"50dfe63c-b3b2-46c5-ad8d-91c64e3ef698","Type":"ContainerStarted","Data":"f8a6c115c16ea56885c902ae0bb872c0ba0d9ec948437bebfa35e773f2629c38"} Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.832092 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.838088 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" event={"ID":"f4fddbf6-0bc5-4b36-9a8a-7073e875c540","Type":"ContainerStarted","Data":"48bdf02edc4f3c05929ef2049c323c31b5ce0c884a259d8859df6abc4ebd632b"} Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.838131 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.850154 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" podStartSLOduration=32.999099049 podStartE2EDuration="36.850135559s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:57:15.969583065 +0000 UTC m=+1097.924944111" lastFinishedPulling="2025-12-04 17:57:19.820619575 +0000 UTC m=+1101.775980621" observedRunningTime="2025-12-04 17:57:22.095578675 +0000 UTC m=+1104.050939721" watchObservedRunningTime="2025-12-04 17:57:22.850135559 +0000 UTC m=+1104.805496605" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.851271 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" podStartSLOduration=3.766729402 podStartE2EDuration="36.851264198s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.302450608 +0000 UTC m=+1070.257811654" lastFinishedPulling="2025-12-04 17:57:21.386985404 +0000 UTC m=+1103.342346450" observedRunningTime="2025-12-04 17:57:22.835589451 +0000 UTC m=+1104.790950507" watchObservedRunningTime="2025-12-04 17:57:22.851264198 +0000 UTC m=+1104.806625244" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.855711 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m45ps" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.855778 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-txjrm" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.855831 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-k5gsb" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.860668 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" podStartSLOduration=3.6777112770000002 podStartE2EDuration="36.860643258s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.136014914 +0000 UTC m=+1070.091375960" lastFinishedPulling="2025-12-04 17:57:21.318946895 +0000 UTC m=+1103.274307941" observedRunningTime="2025-12-04 17:57:22.85843595 +0000 UTC m=+1104.813797006" watchObservedRunningTime="2025-12-04 17:57:22.860643258 +0000 UTC m=+1104.816004324" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.890070 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" podStartSLOduration=3.265130997 podStartE2EDuration="36.89005085s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.443254553 +0000 UTC m=+1070.398615599" lastFinishedPulling="2025-12-04 17:57:22.068174406 +0000 UTC m=+1104.023535452" observedRunningTime="2025-12-04 17:57:22.881347169 +0000 UTC m=+1104.836708215" watchObservedRunningTime="2025-12-04 17:57:22.89005085 +0000 UTC m=+1104.845411896" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.912064 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" podStartSLOduration=2.558695927 podStartE2EDuration="36.912046985s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:47.875399878 +0000 UTC m=+1069.830760924" lastFinishedPulling="2025-12-04 17:57:22.228750936 +0000 UTC m=+1104.184111982" observedRunningTime="2025-12-04 17:57:22.909154048 +0000 UTC m=+1104.864515094" watchObservedRunningTime="2025-12-04 17:57:22.912046985 +0000 UTC m=+1104.867408031" Dec 04 17:57:22 crc kubenswrapper[4733]: I1204 17:57:22.928909 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" podStartSLOduration=3.134523272 podStartE2EDuration="36.928888163s" podCreationTimestamp="2025-12-04 17:56:46 +0000 UTC" firstStartedPulling="2025-12-04 17:56:48.435922806 +0000 UTC m=+1070.391283852" lastFinishedPulling="2025-12-04 17:57:22.230287697 +0000 UTC m=+1104.185648743" observedRunningTime="2025-12-04 17:57:22.925399669 +0000 UTC m=+1104.880760715" watchObservedRunningTime="2025-12-04 17:57:22.928888163 +0000 UTC m=+1104.884249209" Dec 04 17:57:23 crc kubenswrapper[4733]: I1204 17:57:23.608034 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-gblbb" Dec 04 17:57:27 crc kubenswrapper[4733]: I1204 17:57:27.032724 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bjzvv" Dec 04 17:57:27 crc kubenswrapper[4733]: I1204 17:57:27.063652 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-25jzc" Dec 04 17:57:27 crc kubenswrapper[4733]: I1204 17:57:27.171956 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-v8d4h" Dec 04 17:57:27 crc kubenswrapper[4733]: I1204 17:57:27.276663 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qpssp" Dec 04 17:57:27 crc kubenswrapper[4733]: I1204 17:57:27.411279 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-5rln8" Dec 04 17:57:27 crc kubenswrapper[4733]: I1204 17:57:27.564164 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7qmwh" Dec 04 17:57:27 crc kubenswrapper[4733]: I1204 17:57:27.683295 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-7rb9w" Dec 04 17:57:27 crc kubenswrapper[4733]: I1204 17:57:27.972860 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-mf6w7" Dec 04 17:57:32 crc kubenswrapper[4733]: I1204 17:57:32.781943 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-8skx7" Dec 04 17:57:33 crc kubenswrapper[4733]: I1204 17:57:33.082525 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f59cdsz" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.582366 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-5ljmx"] Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.583942 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.586854 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wztgl" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.587618 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.587978 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.588149 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.601950 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-5ljmx"] Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.662618 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d38fee4-7f97-41be-bc98-fdfc32b93f18-config\") pod \"dnsmasq-dns-5cd484bb89-5ljmx\" (UID: \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\") " pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.662732 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb2nx\" (UniqueName: \"kubernetes.io/projected/6d38fee4-7f97-41be-bc98-fdfc32b93f18-kube-api-access-mb2nx\") pod \"dnsmasq-dns-5cd484bb89-5ljmx\" (UID: \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\") " pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.672729 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-567c455747-gqcrm"] Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.677639 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.679253 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.688083 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-gqcrm"] Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.763757 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-config\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.763972 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d38fee4-7f97-41be-bc98-fdfc32b93f18-config\") pod \"dnsmasq-dns-5cd484bb89-5ljmx\" (UID: \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\") " pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.764032 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-dns-svc\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.764084 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jckf4\" (UniqueName: \"kubernetes.io/projected/65b10cc9-360f-463b-9906-bd398bbbdf4e-kube-api-access-jckf4\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.764148 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb2nx\" (UniqueName: \"kubernetes.io/projected/6d38fee4-7f97-41be-bc98-fdfc32b93f18-kube-api-access-mb2nx\") pod \"dnsmasq-dns-5cd484bb89-5ljmx\" (UID: \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\") " pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.781879 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb2nx\" (UniqueName: \"kubernetes.io/projected/6d38fee4-7f97-41be-bc98-fdfc32b93f18-kube-api-access-mb2nx\") pod \"dnsmasq-dns-5cd484bb89-5ljmx\" (UID: \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\") " pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.865201 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-config\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.865278 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-dns-svc\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.865300 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jckf4\" (UniqueName: \"kubernetes.io/projected/65b10cc9-360f-463b-9906-bd398bbbdf4e-kube-api-access-jckf4\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.866088 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-dns-svc\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.867366 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-config\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.887844 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jckf4\" (UniqueName: \"kubernetes.io/projected/65b10cc9-360f-463b-9906-bd398bbbdf4e-kube-api-access-jckf4\") pod \"dnsmasq-dns-567c455747-gqcrm\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:50 crc kubenswrapper[4733]: I1204 17:57:50.995905 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:57:52 crc kubenswrapper[4733]: I1204 17:57:52.507955 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d38fee4-7f97-41be-bc98-fdfc32b93f18-config\") pod \"dnsmasq-dns-5cd484bb89-5ljmx\" (UID: \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\") " pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:57:52 crc kubenswrapper[4733]: I1204 17:57:52.704922 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:57:52 crc kubenswrapper[4733]: I1204 17:57:52.794546 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-gqcrm"] Dec 04 17:57:52 crc kubenswrapper[4733]: I1204 17:57:52.822849 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-gqcrm"] Dec 04 17:57:52 crc kubenswrapper[4733]: I1204 17:57:52.829328 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-859d485f47-zf8cv"] Dec 04 17:57:52 crc kubenswrapper[4733]: I1204 17:57:52.830655 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:52 crc kubenswrapper[4733]: I1204 17:57:52.839843 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-zf8cv"] Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.003163 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-config\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.003204 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-dns-svc\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.003248 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sznvm\" (UniqueName: \"kubernetes.io/projected/01770c66-61c0-4290-9bd3-de725c6814b0-kube-api-access-sznvm\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.095940 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-5ljmx"] Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.106022 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sznvm\" (UniqueName: \"kubernetes.io/projected/01770c66-61c0-4290-9bd3-de725c6814b0-kube-api-access-sznvm\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.106137 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-config\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.106159 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-dns-svc\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.107025 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-dns-svc\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.107540 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-config\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.118401 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-gqcrm" event={"ID":"65b10cc9-360f-463b-9906-bd398bbbdf4e","Type":"ContainerStarted","Data":"e8e7ae36b703e3b0d7eeadd238588943f6f178b14fcbd4cbb807a78ab37f9216"} Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.124581 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb666b895-7vs5b"] Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.125675 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.132633 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sznvm\" (UniqueName: \"kubernetes.io/projected/01770c66-61c0-4290-9bd3-de725c6814b0-kube-api-access-sznvm\") pod \"dnsmasq-dns-859d485f47-zf8cv\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.145052 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-7vs5b"] Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.230031 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.309062 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-dns-svc\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.309148 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-config\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.309193 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vkw9\" (UniqueName: \"kubernetes.io/projected/5d3a2277-a4db-4929-8fa2-67462cdc88fa-kube-api-access-4vkw9\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.391180 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-5ljmx"] Dec 04 17:57:53 crc kubenswrapper[4733]: W1204 17:57:53.400005 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d38fee4_7f97_41be_bc98_fdfc32b93f18.slice/crio-3f892fae124d3c7e8800085fa11ec1354a12891e82c4042600ab8b78fb31d2b9 WatchSource:0}: Error finding container 3f892fae124d3c7e8800085fa11ec1354a12891e82c4042600ab8b78fb31d2b9: Status 404 returned error can't find the container with id 3f892fae124d3c7e8800085fa11ec1354a12891e82c4042600ab8b78fb31d2b9 Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.411442 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vkw9\" (UniqueName: \"kubernetes.io/projected/5d3a2277-a4db-4929-8fa2-67462cdc88fa-kube-api-access-4vkw9\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.411502 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-dns-svc\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.411893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-config\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.412208 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-dns-svc\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.412642 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-config\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.425852 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vkw9\" (UniqueName: \"kubernetes.io/projected/5d3a2277-a4db-4929-8fa2-67462cdc88fa-kube-api-access-4vkw9\") pod \"dnsmasq-dns-cb666b895-7vs5b\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.448302 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.572013 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-zf8cv"] Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.878334 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-7vs5b"] Dec 04 17:57:53 crc kubenswrapper[4733]: W1204 17:57:53.893295 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d3a2277_a4db_4929_8fa2_67462cdc88fa.slice/crio-bf65c11ea771e89e5627101f6c4bbde25f533d98a8a5026b2054b43625f5ac8c WatchSource:0}: Error finding container bf65c11ea771e89e5627101f6c4bbde25f533d98a8a5026b2054b43625f5ac8c: Status 404 returned error can't find the container with id bf65c11ea771e89e5627101f6c4bbde25f533d98a8a5026b2054b43625f5ac8c Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.950683 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.952288 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.955056 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.955509 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.955663 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.955759 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6cn2m" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.956663 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.956813 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 17:57:53 crc kubenswrapper[4733]: I1204 17:57:53.956847 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.053338 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.128872 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.128979 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129058 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e605c62-64b4-4417-80bb-bc3387881f7a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129100 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h65g\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-kube-api-access-8h65g\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129195 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129235 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129331 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129369 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129396 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e605c62-64b4-4417-80bb-bc3387881f7a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129438 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.129482 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.141120 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" event={"ID":"6d38fee4-7f97-41be-bc98-fdfc32b93f18","Type":"ContainerStarted","Data":"3f892fae124d3c7e8800085fa11ec1354a12891e82c4042600ab8b78fb31d2b9"} Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.142496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-7vs5b" event={"ID":"5d3a2277-a4db-4929-8fa2-67462cdc88fa","Type":"ContainerStarted","Data":"bf65c11ea771e89e5627101f6c4bbde25f533d98a8a5026b2054b43625f5ac8c"} Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.143942 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-zf8cv" event={"ID":"01770c66-61c0-4290-9bd3-de725c6814b0","Type":"ContainerStarted","Data":"c36b52a97a48c2ba19b3b9a3cb9e8d10d3fcfc94bf8c931f37d31590dcd73215"} Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.231536 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.231635 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.231666 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e605c62-64b4-4417-80bb-bc3387881f7a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.232259 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.232491 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h65g\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-kube-api-access-8h65g\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.232647 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.232711 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.232836 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.232878 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.232911 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e605c62-64b4-4417-80bb-bc3387881f7a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.232960 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.233014 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.233150 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.234469 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.239780 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.241092 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.242140 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.243031 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.244128 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e605c62-64b4-4417-80bb-bc3387881f7a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.248611 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.255630 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e605c62-64b4-4417-80bb-bc3387881f7a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.265396 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h65g\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-kube-api-access-8h65g\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.267396 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.268995 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.283140 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.285899 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.301312 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.307144 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.307397 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.307546 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.307775 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wrbbz" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.307946 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.308079 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436097 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f55e362-d59c-4269-92c3-d5ca014a2ef1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436143 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436162 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436181 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436212 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f55e362-d59c-4269-92c3-d5ca014a2ef1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436230 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436254 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436291 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm2ds\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-kube-api-access-hm2ds\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436310 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436330 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.436361 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538080 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538170 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538259 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm2ds\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-kube-api-access-hm2ds\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538306 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538366 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538426 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538484 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f55e362-d59c-4269-92c3-d5ca014a2ef1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538513 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538534 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538557 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538593 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f55e362-d59c-4269-92c3-d5ca014a2ef1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.538814 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.539074 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.539325 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.539466 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.539471 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.540283 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.543042 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f55e362-d59c-4269-92c3-d5ca014a2ef1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.551488 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f55e362-d59c-4269-92c3-d5ca014a2ef1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.558898 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.559334 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.560628 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm2ds\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-kube-api-access-hm2ds\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.565461 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " pod="openstack/rabbitmq-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.573221 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:57:54 crc kubenswrapper[4733]: I1204 17:57:54.693022 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.080826 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 17:57:55 crc kubenswrapper[4733]: W1204 17:57:55.086733 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e605c62_64b4_4417_80bb_bc3387881f7a.slice/crio-9081bacac85505f2a714990eec713b64d0c250ae5e824623e7cd9d39326a3676 WatchSource:0}: Error finding container 9081bacac85505f2a714990eec713b64d0c250ae5e824623e7cd9d39326a3676: Status 404 returned error can't find the container with id 9081bacac85505f2a714990eec713b64d0c250ae5e824623e7cd9d39326a3676 Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.155574 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e605c62-64b4-4417-80bb-bc3387881f7a","Type":"ContainerStarted","Data":"9081bacac85505f2a714990eec713b64d0c250ae5e824623e7cd9d39326a3676"} Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.178654 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.546518 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.547745 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.554132 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.554163 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.558922 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.559170 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-c4nlc" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.561902 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.565355 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.656359 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.656408 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.656445 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.656468 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-kolla-config\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.656497 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.656518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hb7p\" (UniqueName: \"kubernetes.io/projected/b9cb46d6-3619-4035-8431-0b8552c5e690-kube-api-access-8hb7p\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.656559 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.656688 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-default\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759441 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759490 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-kolla-config\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759531 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759555 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hb7p\" (UniqueName: \"kubernetes.io/projected/b9cb46d6-3619-4035-8431-0b8552c5e690-kube-api-access-8hb7p\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759599 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759667 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-default\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759711 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.759874 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.760741 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-kolla-config\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.760767 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-default\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.761013 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.762249 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.765648 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.781766 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hb7p\" (UniqueName: \"kubernetes.io/projected/b9cb46d6-3619-4035-8431-0b8552c5e690-kube-api-access-8hb7p\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.784256 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.814371 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " pod="openstack/openstack-galera-0" Dec 04 17:57:55 crc kubenswrapper[4733]: I1204 17:57:55.901490 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 17:57:56 crc kubenswrapper[4733]: I1204 17:57:56.168148 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f55e362-d59c-4269-92c3-d5ca014a2ef1","Type":"ContainerStarted","Data":"53a625c56c3f13329a1fe71e23d20de407a9ea740aa83d404bad031a89416a66"} Dec 04 17:57:56 crc kubenswrapper[4733]: I1204 17:57:56.477298 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 17:57:56 crc kubenswrapper[4733]: W1204 17:57:56.525614 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9cb46d6_3619_4035_8431_0b8552c5e690.slice/crio-c3601ba38f79f631e58366ddf941e47700a1f4d3e180b9a182e1776952c9f6a9 WatchSource:0}: Error finding container c3601ba38f79f631e58366ddf941e47700a1f4d3e180b9a182e1776952c9f6a9: Status 404 returned error can't find the container with id c3601ba38f79f631e58366ddf941e47700a1f4d3e180b9a182e1776952c9f6a9 Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.019644 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.021221 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.033760 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-zjfj7" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.036283 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.039297 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.042056 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.042269 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.087535 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.087624 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.087661 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.087714 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.087751 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.087859 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.087889 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpf66\" (UniqueName: \"kubernetes.io/projected/5833da06-817f-4451-be68-9a9a94e5ef64-kube-api-access-bpf66\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.087932 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189375 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189432 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpf66\" (UniqueName: \"kubernetes.io/projected/5833da06-817f-4451-be68-9a9a94e5ef64-kube-api-access-bpf66\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189466 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189501 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189552 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189588 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189614 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189649 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.189962 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.190227 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.190562 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.191656 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.194784 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.198722 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.199819 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9cb46d6-3619-4035-8431-0b8552c5e690","Type":"ContainerStarted","Data":"c3601ba38f79f631e58366ddf941e47700a1f4d3e180b9a182e1776952c9f6a9"} Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.212739 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.232776 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpf66\" (UniqueName: \"kubernetes.io/projected/5833da06-817f-4451-be68-9a9a94e5ef64-kube-api-access-bpf66\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.239071 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.339085 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.341180 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.343669 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.344424 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.344642 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-cs5hs" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.360044 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.367863 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.394922 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-kolla-config\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.394970 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.395104 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gd8p\" (UniqueName: \"kubernetes.io/projected/3d28f771-2fba-42d9-8fc0-1107d17db527-kube-api-access-9gd8p\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.395154 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-config-data\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.395177 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.496890 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gd8p\" (UniqueName: \"kubernetes.io/projected/3d28f771-2fba-42d9-8fc0-1107d17db527-kube-api-access-9gd8p\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.496946 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-config-data\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.496974 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.497023 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-kolla-config\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.497088 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.500699 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-config-data\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.501498 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-kolla-config\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.513870 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.522343 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.535933 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gd8p\" (UniqueName: \"kubernetes.io/projected/3d28f771-2fba-42d9-8fc0-1107d17db527-kube-api-access-9gd8p\") pod \"memcached-0\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " pod="openstack/memcached-0" Dec 04 17:57:57 crc kubenswrapper[4733]: I1204 17:57:57.680427 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 17:57:58 crc kubenswrapper[4733]: I1204 17:57:58.120887 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 17:57:58 crc kubenswrapper[4733]: W1204 17:57:58.138904 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5833da06_817f_4451_be68_9a9a94e5ef64.slice/crio-500dfcc1ae47b8d3bd343720e7a4af1ba091cdd3ba2dd568851fc0a4057d4c17 WatchSource:0}: Error finding container 500dfcc1ae47b8d3bd343720e7a4af1ba091cdd3ba2dd568851fc0a4057d4c17: Status 404 returned error can't find the container with id 500dfcc1ae47b8d3bd343720e7a4af1ba091cdd3ba2dd568851fc0a4057d4c17 Dec 04 17:57:58 crc kubenswrapper[4733]: I1204 17:57:58.209237 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5833da06-817f-4451-be68-9a9a94e5ef64","Type":"ContainerStarted","Data":"500dfcc1ae47b8d3bd343720e7a4af1ba091cdd3ba2dd568851fc0a4057d4c17"} Dec 04 17:57:58 crc kubenswrapper[4733]: I1204 17:57:58.284238 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.226001 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3d28f771-2fba-42d9-8fc0-1107d17db527","Type":"ContainerStarted","Data":"f9c96f88c3e7c60ff1d0226643539b80718d7ba96846aeb939ea4ed7c48fa513"} Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.476090 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.477307 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.479578 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-8zct5" Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.490925 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.589838 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbsgv\" (UniqueName: \"kubernetes.io/projected/47e99629-3bc7-4756-9b65-11ab755f49ad-kube-api-access-fbsgv\") pod \"kube-state-metrics-0\" (UID: \"47e99629-3bc7-4756-9b65-11ab755f49ad\") " pod="openstack/kube-state-metrics-0" Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.691653 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbsgv\" (UniqueName: \"kubernetes.io/projected/47e99629-3bc7-4756-9b65-11ab755f49ad-kube-api-access-fbsgv\") pod \"kube-state-metrics-0\" (UID: \"47e99629-3bc7-4756-9b65-11ab755f49ad\") " pod="openstack/kube-state-metrics-0" Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.722072 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbsgv\" (UniqueName: \"kubernetes.io/projected/47e99629-3bc7-4756-9b65-11ab755f49ad-kube-api-access-fbsgv\") pod \"kube-state-metrics-0\" (UID: \"47e99629-3bc7-4756-9b65-11ab755f49ad\") " pod="openstack/kube-state-metrics-0" Dec 04 17:57:59 crc kubenswrapper[4733]: I1204 17:57:59.811379 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 17:58:00 crc kubenswrapper[4733]: W1204 17:58:00.368684 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47e99629_3bc7_4756_9b65_11ab755f49ad.slice/crio-680494eef17cf2b64057b1af3d5744b0d26f565905768805633001a6fbecc5b9 WatchSource:0}: Error finding container 680494eef17cf2b64057b1af3d5744b0d26f565905768805633001a6fbecc5b9: Status 404 returned error can't find the container with id 680494eef17cf2b64057b1af3d5744b0d26f565905768805633001a6fbecc5b9 Dec 04 17:58:00 crc kubenswrapper[4733]: I1204 17:58:00.371650 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 17:58:01 crc kubenswrapper[4733]: I1204 17:58:01.247701 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"47e99629-3bc7-4756-9b65-11ab755f49ad","Type":"ContainerStarted","Data":"680494eef17cf2b64057b1af3d5744b0d26f565905768805633001a6fbecc5b9"} Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.803762 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-j58xt"] Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.804921 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.807919 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-kvx56" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.808301 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.810417 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.826870 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-j58xt"] Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.863003 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.863047 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run-ovn\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.863128 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j46r9\" (UniqueName: \"kubernetes.io/projected/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-kube-api-access-j46r9\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.863179 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-scripts\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.863202 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-combined-ca-bundle\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.863239 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-ovn-controller-tls-certs\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.863293 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-log-ovn\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.882955 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-rkmmv"] Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.886931 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.894443 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rkmmv"] Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.964571 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.964625 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run-ovn\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965120 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965169 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run-ovn\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.964690 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-lib\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965257 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-run\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965403 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-scripts\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965498 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rprg6\" (UniqueName: \"kubernetes.io/projected/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-kube-api-access-rprg6\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965580 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j46r9\" (UniqueName: \"kubernetes.io/projected/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-kube-api-access-j46r9\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965613 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-etc-ovs\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965643 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-log\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.965836 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-scripts\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.968336 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-scripts\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.968489 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-combined-ca-bundle\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.968546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-ovn-controller-tls-certs\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.968592 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-log-ovn\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.968877 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-log-ovn\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.986624 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-ovn-controller-tls-certs\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.987500 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-combined-ca-bundle\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:03 crc kubenswrapper[4733]: I1204 17:58:03.988857 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j46r9\" (UniqueName: \"kubernetes.io/projected/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-kube-api-access-j46r9\") pod \"ovn-controller-j58xt\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " pod="openstack/ovn-controller-j58xt" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.046840 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.048047 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.053164 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-x7cnc" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.056291 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.056345 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.056485 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.056889 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.070330 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-etc-ovs\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.070376 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-log\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.070450 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-lib\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.070469 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-run\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.070496 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-scripts\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.070522 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rprg6\" (UniqueName: \"kubernetes.io/projected/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-kube-api-access-rprg6\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.071830 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-etc-ovs\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.071908 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-log\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.072000 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-lib\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.072037 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-run\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.073659 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-scripts\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.079220 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.091464 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rprg6\" (UniqueName: \"kubernetes.io/projected/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-kube-api-access-rprg6\") pod \"ovn-controller-ovs-rkmmv\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.133119 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.171719 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.171778 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghkf6\" (UniqueName: \"kubernetes.io/projected/bae2e2ab-b6af-4d4d-8288-9f23050553a0-kube-api-access-ghkf6\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.171827 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.171850 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.171877 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.171911 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.171938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.171954 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.202626 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.273068 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.273121 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.273151 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.273184 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.273215 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.273332 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghkf6\" (UniqueName: \"kubernetes.io/projected/bae2e2ab-b6af-4d4d-8288-9f23050553a0-kube-api-access-ghkf6\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.273365 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.274091 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.274087 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.274157 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.274456 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.273384 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.276428 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.277400 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.277667 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.292874 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghkf6\" (UniqueName: \"kubernetes.io/projected/bae2e2ab-b6af-4d4d-8288-9f23050553a0-kube-api-access-ghkf6\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.315986 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:04 crc kubenswrapper[4733]: I1204 17:58:04.380600 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.175763 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.177399 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.183616 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.228190 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.228433 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-hzjhw" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.228470 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.228575 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.331203 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.331281 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.331320 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.331379 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.331401 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpqgh\" (UniqueName: \"kubernetes.io/projected/ae61093e-d566-48cb-acc2-dd88c4512b8e-kube-api-access-fpqgh\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.331466 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.331488 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.331514 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-config\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.432881 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.433264 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.433298 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.433355 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.433380 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpqgh\" (UniqueName: \"kubernetes.io/projected/ae61093e-d566-48cb-acc2-dd88c4512b8e-kube-api-access-fpqgh\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.433439 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.433461 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.433485 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-config\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.433654 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.434236 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-config\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.434244 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.434291 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.440679 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.442081 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.446053 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.462623 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpqgh\" (UniqueName: \"kubernetes.io/projected/ae61093e-d566-48cb-acc2-dd88c4512b8e-kube-api-access-fpqgh\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.501823 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:06 crc kubenswrapper[4733]: I1204 17:58:06.545138 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:15 crc kubenswrapper[4733]: I1204 17:58:15.361876 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:58:15 crc kubenswrapper[4733]: I1204 17:58:15.362385 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:58:29 crc kubenswrapper[4733]: E1204 17:58:29.494593 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 04 17:58:29 crc kubenswrapper[4733]: E1204 17:58:29.495252 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8hb7p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(b9cb46d6-3619-4035-8431-0b8552c5e690): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:29 crc kubenswrapper[4733]: E1204 17:58:29.497074 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" Dec 04 17:58:29 crc kubenswrapper[4733]: E1204 17:58:29.520182 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-galera-0" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" Dec 04 17:58:30 crc kubenswrapper[4733]: E1204 17:58:30.704642 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 04 17:58:30 crc kubenswrapper[4733]: E1204 17:58:30.705174 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8h65g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(5e605c62-64b4-4417-80bb-bc3387881f7a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:30 crc kubenswrapper[4733]: E1204 17:58:30.706493 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.485385 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.485590 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n7bh79h5dfh67dh5bdh78h5h5f7h564h67hdchd7h5f9h678h5bh5fhbfh597h594h5f9h56fh54h5bch5ffh68h6dh556h4h54bh57bh5dch57cq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gd8p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(3d28f771-2fba-42d9-8fc0-1107d17db527): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.486762 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="3d28f771-2fba-42d9-8fc0-1107d17db527" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.523661 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.524870 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hm2ds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(1f55e362-d59c-4269-92c3-d5ca014a2ef1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.526051 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.550425 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.550599 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108\\\"\"" pod="openstack/memcached-0" podUID="3d28f771-2fba-42d9-8fc0-1107d17db527" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.587155 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.587332 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bpf66,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(5833da06-817f-4451-be68-9a9a94e5ef64): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:31 crc kubenswrapper[4733]: E1204 17:58:31.588654 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.356031 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.356095 4733 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.356239 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fbsgv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(47e99629-3bc7-4756-9b65-11ab755f49ad): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" logger="UnhandledError" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.357358 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="47e99629-3bc7-4756-9b65-11ab755f49ad" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.363143 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.363647 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jckf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-567c455747-gqcrm_openstack(65b10cc9-360f-463b-9906-bd398bbbdf4e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.365193 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-567c455747-gqcrm" podUID="65b10cc9-360f-463b-9906-bd398bbbdf4e" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.380989 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.381169 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sznvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-859d485f47-zf8cv_openstack(01770c66-61c0-4290-9bd3-de725c6814b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.382221 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.382305 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4vkw9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-cb666b895-7vs5b_openstack(5d3a2277-a4db-4929-8fa2-67462cdc88fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.382349 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-859d485f47-zf8cv" podUID="01770c66-61c0-4290-9bd3-de725c6814b0" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.383841 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-cb666b895-7vs5b" podUID="5d3a2277-a4db-4929-8fa2-67462cdc88fa" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.388420 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.388661 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mb2nx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5cd484bb89-5ljmx_openstack(6d38fee4-7f97-41be-bc98-fdfc32b93f18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.389869 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" podUID="6d38fee4-7f97-41be-bc98-fdfc32b93f18" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.569566 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-cb666b895-7vs5b" podUID="5d3a2277-a4db-4929-8fa2-67462cdc88fa" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.570137 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-859d485f47-zf8cv" podUID="01770c66-61c0-4290-9bd3-de725c6814b0" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.570677 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-server-0" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.570468 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" Dec 04 17:58:32 crc kubenswrapper[4733]: E1204 17:58:32.572547 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="47e99629-3bc7-4756-9b65-11ab755f49ad" Dec 04 17:58:32 crc kubenswrapper[4733]: I1204 17:58:32.959013 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-j58xt"] Dec 04 17:58:32 crc kubenswrapper[4733]: I1204 17:58:32.965382 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:58:32 crc kubenswrapper[4733]: I1204 17:58:32.992308 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.047350 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.099675 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d38fee4-7f97-41be-bc98-fdfc32b93f18-config\") pod \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\" (UID: \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\") " Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.099767 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jckf4\" (UniqueName: \"kubernetes.io/projected/65b10cc9-360f-463b-9906-bd398bbbdf4e-kube-api-access-jckf4\") pod \"65b10cc9-360f-463b-9906-bd398bbbdf4e\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.099787 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb2nx\" (UniqueName: \"kubernetes.io/projected/6d38fee4-7f97-41be-bc98-fdfc32b93f18-kube-api-access-mb2nx\") pod \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\" (UID: \"6d38fee4-7f97-41be-bc98-fdfc32b93f18\") " Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.099828 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-dns-svc\") pod \"65b10cc9-360f-463b-9906-bd398bbbdf4e\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.099875 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-config\") pod \"65b10cc9-360f-463b-9906-bd398bbbdf4e\" (UID: \"65b10cc9-360f-463b-9906-bd398bbbdf4e\") " Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.101462 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-config" (OuterVolumeSpecName: "config") pod "65b10cc9-360f-463b-9906-bd398bbbdf4e" (UID: "65b10cc9-360f-463b-9906-bd398bbbdf4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.101705 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d38fee4-7f97-41be-bc98-fdfc32b93f18-config" (OuterVolumeSpecName: "config") pod "6d38fee4-7f97-41be-bc98-fdfc32b93f18" (UID: "6d38fee4-7f97-41be-bc98-fdfc32b93f18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.103940 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "65b10cc9-360f-463b-9906-bd398bbbdf4e" (UID: "65b10cc9-360f-463b-9906-bd398bbbdf4e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.106829 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b10cc9-360f-463b-9906-bd398bbbdf4e-kube-api-access-jckf4" (OuterVolumeSpecName: "kube-api-access-jckf4") pod "65b10cc9-360f-463b-9906-bd398bbbdf4e" (UID: "65b10cc9-360f-463b-9906-bd398bbbdf4e"). InnerVolumeSpecName "kube-api-access-jckf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.106916 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d38fee4-7f97-41be-bc98-fdfc32b93f18-kube-api-access-mb2nx" (OuterVolumeSpecName: "kube-api-access-mb2nx") pod "6d38fee4-7f97-41be-bc98-fdfc32b93f18" (UID: "6d38fee4-7f97-41be-bc98-fdfc32b93f18"). InnerVolumeSpecName "kube-api-access-mb2nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.150620 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rkmmv"] Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.201475 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d38fee4-7f97-41be-bc98-fdfc32b93f18-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.201529 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jckf4\" (UniqueName: \"kubernetes.io/projected/65b10cc9-360f-463b-9906-bd398bbbdf4e-kube-api-access-jckf4\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.201551 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb2nx\" (UniqueName: \"kubernetes.io/projected/6d38fee4-7f97-41be-bc98-fdfc32b93f18-kube-api-access-mb2nx\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.201570 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.201587 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65b10cc9-360f-463b-9906-bd398bbbdf4e-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.575665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bae2e2ab-b6af-4d4d-8288-9f23050553a0","Type":"ContainerStarted","Data":"f70927bc9d53b516e4d204c03b0fb379b1c49f82a18c06f3a57df744e6f37932"} Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.578260 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.578297 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-5ljmx" event={"ID":"6d38fee4-7f97-41be-bc98-fdfc32b93f18","Type":"ContainerDied","Data":"3f892fae124d3c7e8800085fa11ec1354a12891e82c4042600ab8b78fb31d2b9"} Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.581554 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-gqcrm" event={"ID":"65b10cc9-360f-463b-9906-bd398bbbdf4e","Type":"ContainerDied","Data":"e8e7ae36b703e3b0d7eeadd238588943f6f178b14fcbd4cbb807a78ab37f9216"} Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.581599 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-gqcrm" Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.584976 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt" event={"ID":"3accb1a1-6ad9-42b9-9894-ba7b9e01d919","Type":"ContainerStarted","Data":"46ddd1e4cb00852ea370b1849f0d03314b82549b44b2ce83ff01416b7b9abc9a"} Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.586787 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rkmmv" event={"ID":"0cf011d5-1c33-4ec6-8b1f-d1216a46d066","Type":"ContainerStarted","Data":"ad730f55d153bbc15f9343dc147daeed6c5833d75283b532b5091379e2f35d53"} Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.678564 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-5ljmx"] Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.685259 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-5ljmx"] Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.717853 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-gqcrm"] Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.742024 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-567c455747-gqcrm"] Dec 04 17:58:33 crc kubenswrapper[4733]: I1204 17:58:33.954587 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 17:58:33 crc kubenswrapper[4733]: W1204 17:58:33.958729 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae61093e_d566_48cb_acc2_dd88c4512b8e.slice/crio-162cddb3f0e641814c81d2b1c26c22e815c3678b6df38c946878e1635dfee126 WatchSource:0}: Error finding container 162cddb3f0e641814c81d2b1c26c22e815c3678b6df38c946878e1635dfee126: Status 404 returned error can't find the container with id 162cddb3f0e641814c81d2b1c26c22e815c3678b6df38c946878e1635dfee126 Dec 04 17:58:34 crc kubenswrapper[4733]: I1204 17:58:34.350415 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65b10cc9-360f-463b-9906-bd398bbbdf4e" path="/var/lib/kubelet/pods/65b10cc9-360f-463b-9906-bd398bbbdf4e/volumes" Dec 04 17:58:34 crc kubenswrapper[4733]: I1204 17:58:34.351285 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d38fee4-7f97-41be-bc98-fdfc32b93f18" path="/var/lib/kubelet/pods/6d38fee4-7f97-41be-bc98-fdfc32b93f18/volumes" Dec 04 17:58:34 crc kubenswrapper[4733]: I1204 17:58:34.595415 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ae61093e-d566-48cb-acc2-dd88c4512b8e","Type":"ContainerStarted","Data":"162cddb3f0e641814c81d2b1c26c22e815c3678b6df38c946878e1635dfee126"} Dec 04 17:58:37 crc kubenswrapper[4733]: I1204 17:58:37.620062 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ae61093e-d566-48cb-acc2-dd88c4512b8e","Type":"ContainerStarted","Data":"f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052"} Dec 04 17:58:37 crc kubenswrapper[4733]: I1204 17:58:37.621975 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt" event={"ID":"3accb1a1-6ad9-42b9-9894-ba7b9e01d919","Type":"ContainerStarted","Data":"ab4787c533c2ddd15a894f803c59aaec9e2a588d19932dfde226bd69bf3dff66"} Dec 04 17:58:37 crc kubenswrapper[4733]: I1204 17:58:37.622118 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-j58xt" Dec 04 17:58:37 crc kubenswrapper[4733]: I1204 17:58:37.624659 4733 generic.go:334] "Generic (PLEG): container finished" podID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerID="18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc" exitCode=0 Dec 04 17:58:37 crc kubenswrapper[4733]: I1204 17:58:37.624712 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rkmmv" event={"ID":"0cf011d5-1c33-4ec6-8b1f-d1216a46d066","Type":"ContainerDied","Data":"18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc"} Dec 04 17:58:37 crc kubenswrapper[4733]: I1204 17:58:37.626607 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bae2e2ab-b6af-4d4d-8288-9f23050553a0","Type":"ContainerStarted","Data":"995e1f7fa4114b63bde0864f52338990697384923eb1c3ee3773a27841eb999d"} Dec 04 17:58:37 crc kubenswrapper[4733]: I1204 17:58:37.640554 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-j58xt" podStartSLOduration=30.933364061 podStartE2EDuration="34.640539873s" podCreationTimestamp="2025-12-04 17:58:03 +0000 UTC" firstStartedPulling="2025-12-04 17:58:32.99045497 +0000 UTC m=+1174.945816016" lastFinishedPulling="2025-12-04 17:58:36.697630782 +0000 UTC m=+1178.652991828" observedRunningTime="2025-12-04 17:58:37.636877466 +0000 UTC m=+1179.592238522" watchObservedRunningTime="2025-12-04 17:58:37.640539873 +0000 UTC m=+1179.595900919" Dec 04 17:58:38 crc kubenswrapper[4733]: I1204 17:58:38.637887 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rkmmv" event={"ID":"0cf011d5-1c33-4ec6-8b1f-d1216a46d066","Type":"ContainerStarted","Data":"8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1"} Dec 04 17:58:38 crc kubenswrapper[4733]: I1204 17:58:38.638267 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:38 crc kubenswrapper[4733]: I1204 17:58:38.638280 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rkmmv" event={"ID":"0cf011d5-1c33-4ec6-8b1f-d1216a46d066","Type":"ContainerStarted","Data":"102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16"} Dec 04 17:58:38 crc kubenswrapper[4733]: I1204 17:58:38.638294 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:58:38 crc kubenswrapper[4733]: I1204 17:58:38.659958 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-rkmmv" podStartSLOduration=32.126453555 podStartE2EDuration="35.659925598s" podCreationTimestamp="2025-12-04 17:58:03 +0000 UTC" firstStartedPulling="2025-12-04 17:58:33.160227374 +0000 UTC m=+1175.115588420" lastFinishedPulling="2025-12-04 17:58:36.693699427 +0000 UTC m=+1178.649060463" observedRunningTime="2025-12-04 17:58:38.655972083 +0000 UTC m=+1180.611333139" watchObservedRunningTime="2025-12-04 17:58:38.659925598 +0000 UTC m=+1180.615286644" Dec 04 17:58:40 crc kubenswrapper[4733]: I1204 17:58:40.678722 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bae2e2ab-b6af-4d4d-8288-9f23050553a0","Type":"ContainerStarted","Data":"5c2bc4e3e18cd14bdc17ea6d5ff14205f792f1f10c8fef16751c74cb1ffb2a1e"} Dec 04 17:58:40 crc kubenswrapper[4733]: I1204 17:58:40.682297 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ae61093e-d566-48cb-acc2-dd88c4512b8e","Type":"ContainerStarted","Data":"a2f20b96c0e76358892a2e4c8789ed0cb1a66ac87763e230752a2979798603be"} Dec 04 17:58:40 crc kubenswrapper[4733]: I1204 17:58:40.720379 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=30.692397024999998 podStartE2EDuration="37.720357194s" podCreationTimestamp="2025-12-04 17:58:03 +0000 UTC" firstStartedPulling="2025-12-04 17:58:33.053211229 +0000 UTC m=+1175.008572275" lastFinishedPulling="2025-12-04 17:58:40.081171398 +0000 UTC m=+1182.036532444" observedRunningTime="2025-12-04 17:58:40.709098365 +0000 UTC m=+1182.664459501" watchObservedRunningTime="2025-12-04 17:58:40.720357194 +0000 UTC m=+1182.675718250" Dec 04 17:58:40 crc kubenswrapper[4733]: I1204 17:58:40.766572 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=29.665757506 podStartE2EDuration="35.765789112s" podCreationTimestamp="2025-12-04 17:58:05 +0000 UTC" firstStartedPulling="2025-12-04 17:58:33.961005027 +0000 UTC m=+1175.916366073" lastFinishedPulling="2025-12-04 17:58:40.061036633 +0000 UTC m=+1182.016397679" observedRunningTime="2025-12-04 17:58:40.754973944 +0000 UTC m=+1182.710335060" watchObservedRunningTime="2025-12-04 17:58:40.765789112 +0000 UTC m=+1182.721150188" Dec 04 17:58:41 crc kubenswrapper[4733]: I1204 17:58:41.546583 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:42 crc kubenswrapper[4733]: I1204 17:58:42.546191 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:42 crc kubenswrapper[4733]: I1204 17:58:42.609738 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:42 crc kubenswrapper[4733]: I1204 17:58:42.753627 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.121839 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-zf8cv"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.183264 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-dntx9"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.193770 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-67rp6"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.194515 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.194532 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.198232 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-dntx9"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.202214 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.202364 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.205859 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-67rp6"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379127 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379174 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovn-rundir\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379199 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379225 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf38da41-b6c0-4551-9ce9-32146ebc2160-config\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379252 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-combined-ca-bundle\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379272 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-config\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379294 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovs-rundir\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379327 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl4mv\" (UniqueName: \"kubernetes.io/projected/bf38da41-b6c0-4551-9ce9-32146ebc2160-kube-api-access-tl4mv\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379367 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64562\" (UniqueName: \"kubernetes.io/projected/e659c44b-3009-4c5a-8a49-2d338d0d4c37-kube-api-access-64562\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.379408 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-dns-svc\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.381159 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.428279 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-7vs5b"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.480725 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovn-rundir\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.480833 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.480865 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf38da41-b6c0-4551-9ce9-32146ebc2160-config\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.480934 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-combined-ca-bundle\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.480962 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-config\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.480999 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovs-rundir\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.481067 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl4mv\" (UniqueName: \"kubernetes.io/projected/bf38da41-b6c0-4551-9ce9-32146ebc2160-kube-api-access-tl4mv\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.481161 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64562\" (UniqueName: \"kubernetes.io/projected/e659c44b-3009-4c5a-8a49-2d338d0d4c37-kube-api-access-64562\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.481227 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-dns-svc\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.481267 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.481855 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-rg88m"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.483170 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.484598 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-config\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.484841 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovs-rundir\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.489113 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.492126 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovn-rundir\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.493834 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-rg88m"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.496251 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.498934 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.499263 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-dns-svc\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.509310 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-combined-ca-bundle\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.512321 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf38da41-b6c0-4551-9ce9-32146ebc2160-config\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.524762 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl4mv\" (UniqueName: \"kubernetes.io/projected/bf38da41-b6c0-4551-9ce9-32146ebc2160-kube-api-access-tl4mv\") pod \"ovn-controller-metrics-67rp6\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.531169 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-67rp6" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.532078 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64562\" (UniqueName: \"kubernetes.io/projected/e659c44b-3009-4c5a-8a49-2d338d0d4c37-kube-api-access-64562\") pod \"dnsmasq-dns-7cb7d9565c-dntx9\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.562508 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.582516 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.582826 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.582848 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-dns-svc\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.582885 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-config\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.582938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7qs5\" (UniqueName: \"kubernetes.io/projected/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-kube-api-access-b7qs5\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.613960 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.684569 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-config\") pod \"01770c66-61c0-4290-9bd3-de725c6814b0\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.684626 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-dns-svc\") pod \"01770c66-61c0-4290-9bd3-de725c6814b0\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.684650 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sznvm\" (UniqueName: \"kubernetes.io/projected/01770c66-61c0-4290-9bd3-de725c6814b0-kube-api-access-sznvm\") pod \"01770c66-61c0-4290-9bd3-de725c6814b0\" (UID: \"01770c66-61c0-4290-9bd3-de725c6814b0\") " Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.684941 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.684996 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.685019 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-dns-svc\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.685044 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-config\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.685088 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7qs5\" (UniqueName: \"kubernetes.io/projected/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-kube-api-access-b7qs5\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.685249 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-config" (OuterVolumeSpecName: "config") pod "01770c66-61c0-4290-9bd3-de725c6814b0" (UID: "01770c66-61c0-4290-9bd3-de725c6814b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.685814 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.686141 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "01770c66-61c0-4290-9bd3-de725c6814b0" (UID: "01770c66-61c0-4290-9bd3-de725c6814b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.687626 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.687963 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-config\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.689974 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-dns-svc\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.692764 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01770c66-61c0-4290-9bd3-de725c6814b0-kube-api-access-sznvm" (OuterVolumeSpecName: "kube-api-access-sznvm") pod "01770c66-61c0-4290-9bd3-de725c6814b0" (UID: "01770c66-61c0-4290-9bd3-de725c6814b0"). InnerVolumeSpecName "kube-api-access-sznvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.702584 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7qs5\" (UniqueName: \"kubernetes.io/projected/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-kube-api-access-b7qs5\") pod \"dnsmasq-dns-db7757ddc-rg88m\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.708722 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-859d485f47-zf8cv" event={"ID":"01770c66-61c0-4290-9bd3-de725c6814b0","Type":"ContainerDied","Data":"c36b52a97a48c2ba19b3b9a3cb9e8d10d3fcfc94bf8c931f37d31590dcd73215"} Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.708747 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-859d485f47-zf8cv" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.710388 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.753369 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.785092 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-zf8cv"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.812109 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.812153 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01770c66-61c0-4290-9bd3-de725c6814b0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.812165 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sznvm\" (UniqueName: \"kubernetes.io/projected/01770c66-61c0-4290-9bd3-de725c6814b0-kube-api-access-sznvm\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.824435 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-859d485f47-zf8cv"] Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.825926 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.875819 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:58:43 crc kubenswrapper[4733]: I1204 17:58:43.928516 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.001743 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.003341 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.006454 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.006508 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.006664 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.006732 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pjdhp" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.014932 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.017782 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-dns-svc\") pod \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.017928 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vkw9\" (UniqueName: \"kubernetes.io/projected/5d3a2277-a4db-4929-8fa2-67462cdc88fa-kube-api-access-4vkw9\") pod \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.017967 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-config\") pod \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\" (UID: \"5d3a2277-a4db-4929-8fa2-67462cdc88fa\") " Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.018934 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5d3a2277-a4db-4929-8fa2-67462cdc88fa" (UID: "5d3a2277-a4db-4929-8fa2-67462cdc88fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.021215 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d3a2277-a4db-4929-8fa2-67462cdc88fa-kube-api-access-4vkw9" (OuterVolumeSpecName: "kube-api-access-4vkw9") pod "5d3a2277-a4db-4929-8fa2-67462cdc88fa" (UID: "5d3a2277-a4db-4929-8fa2-67462cdc88fa"). InnerVolumeSpecName "kube-api-access-4vkw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.028128 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-config" (OuterVolumeSpecName: "config") pod "5d3a2277-a4db-4929-8fa2-67462cdc88fa" (UID: "5d3a2277-a4db-4929-8fa2-67462cdc88fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.081500 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-67rp6"] Dec 04 17:58:44 crc kubenswrapper[4733]: W1204 17:58:44.091081 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf38da41_b6c0_4551_9ce9_32146ebc2160.slice/crio-be3a9bbf8278f801466aea9f70acb74e4cc5979dd71988bb769e38914b206d82 WatchSource:0}: Error finding container be3a9bbf8278f801466aea9f70acb74e4cc5979dd71988bb769e38914b206d82: Status 404 returned error can't find the container with id be3a9bbf8278f801466aea9f70acb74e4cc5979dd71988bb769e38914b206d82 Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119200 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119236 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7sn8\" (UniqueName: \"kubernetes.io/projected/229fa5f0-e006-4c04-8476-23a7842d292b-kube-api-access-r7sn8\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119261 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-scripts\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119277 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119293 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119334 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-config\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119363 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119438 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119449 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vkw9\" (UniqueName: \"kubernetes.io/projected/5d3a2277-a4db-4929-8fa2-67462cdc88fa-kube-api-access-4vkw9\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.119459 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d3a2277-a4db-4929-8fa2-67462cdc88fa-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.220875 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.220923 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7sn8\" (UniqueName: \"kubernetes.io/projected/229fa5f0-e006-4c04-8476-23a7842d292b-kube-api-access-r7sn8\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.220947 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-scripts\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.220963 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.220977 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.221018 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-config\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.221046 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.221891 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.222633 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-scripts\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.222696 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-config\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.225460 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.226213 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.226220 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.245753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7sn8\" (UniqueName: \"kubernetes.io/projected/229fa5f0-e006-4c04-8476-23a7842d292b-kube-api-access-r7sn8\") pod \"ovn-northd-0\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.322144 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.327865 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-dntx9"] Dec 04 17:58:44 crc kubenswrapper[4733]: W1204 17:58:44.341311 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode659c44b_3009_4c5a_8a49_2d338d0d4c37.slice/crio-c595dd82e504a002be34ad6ca1822cafd965d7b192a8a2c7d143b24e175f0dab WatchSource:0}: Error finding container c595dd82e504a002be34ad6ca1822cafd965d7b192a8a2c7d143b24e175f0dab: Status 404 returned error can't find the container with id c595dd82e504a002be34ad6ca1822cafd965d7b192a8a2c7d143b24e175f0dab Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.348976 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01770c66-61c0-4290-9bd3-de725c6814b0" path="/var/lib/kubelet/pods/01770c66-61c0-4290-9bd3-de725c6814b0/volumes" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.417929 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-rg88m"] Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.717720 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" event={"ID":"a35bc564-4823-4bb3-9a04-7e1bdcc56a17","Type":"ContainerStarted","Data":"d2ca99acbf3e131ee876c40e3c794c60654fc8ca44f088b7896f57e01adb6e1e"} Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.719773 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-7vs5b" event={"ID":"5d3a2277-a4db-4929-8fa2-67462cdc88fa","Type":"ContainerDied","Data":"bf65c11ea771e89e5627101f6c4bbde25f533d98a8a5026b2054b43625f5ac8c"} Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.719879 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-7vs5b" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.722719 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-67rp6" event={"ID":"bf38da41-b6c0-4551-9ce9-32146ebc2160","Type":"ContainerStarted","Data":"afe99239cae679354f70187888627dd752b6e0f56cc59225d53f0c36a6b659f8"} Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.722761 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-67rp6" event={"ID":"bf38da41-b6c0-4551-9ce9-32146ebc2160","Type":"ContainerStarted","Data":"be3a9bbf8278f801466aea9f70acb74e4cc5979dd71988bb769e38914b206d82"} Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.727091 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9cb46d6-3619-4035-8431-0b8552c5e690","Type":"ContainerStarted","Data":"f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6"} Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.729864 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" event={"ID":"e659c44b-3009-4c5a-8a49-2d338d0d4c37","Type":"ContainerStarted","Data":"c595dd82e504a002be34ad6ca1822cafd965d7b192a8a2c7d143b24e175f0dab"} Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.750621 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-67rp6" podStartSLOduration=1.750602856 podStartE2EDuration="1.750602856s" podCreationTimestamp="2025-12-04 17:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:58:44.742423408 +0000 UTC m=+1186.697784464" watchObservedRunningTime="2025-12-04 17:58:44.750602856 +0000 UTC m=+1186.705963902" Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.761674 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.797002 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-7vs5b"] Dec 04 17:58:44 crc kubenswrapper[4733]: I1204 17:58:44.799129 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-7vs5b"] Dec 04 17:58:45 crc kubenswrapper[4733]: I1204 17:58:45.361494 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:58:45 crc kubenswrapper[4733]: I1204 17:58:45.361869 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:58:45 crc kubenswrapper[4733]: I1204 17:58:45.740378 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5833da06-817f-4451-be68-9a9a94e5ef64","Type":"ContainerStarted","Data":"9ccae8be73d6c45906bab0a51e77d0e62fe11a561821293fc12bd3c1b094fd50"} Dec 04 17:58:45 crc kubenswrapper[4733]: I1204 17:58:45.744389 4733 generic.go:334] "Generic (PLEG): container finished" podID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" containerID="8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb" exitCode=0 Dec 04 17:58:45 crc kubenswrapper[4733]: I1204 17:58:45.744479 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" event={"ID":"e659c44b-3009-4c5a-8a49-2d338d0d4c37","Type":"ContainerDied","Data":"8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb"} Dec 04 17:58:45 crc kubenswrapper[4733]: I1204 17:58:45.746305 4733 generic.go:334] "Generic (PLEG): container finished" podID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerID="b7d68bd6f9f6d1d744eb35792427c56d354195501d227587210172786e20992a" exitCode=0 Dec 04 17:58:45 crc kubenswrapper[4733]: I1204 17:58:45.746344 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" event={"ID":"a35bc564-4823-4bb3-9a04-7e1bdcc56a17","Type":"ContainerDied","Data":"b7d68bd6f9f6d1d744eb35792427c56d354195501d227587210172786e20992a"} Dec 04 17:58:45 crc kubenswrapper[4733]: I1204 17:58:45.748249 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"229fa5f0-e006-4c04-8476-23a7842d292b","Type":"ContainerStarted","Data":"aa67deef74c1ebb3aa5e2511dbccc8d8adfecb54e13529613f7271be932a343d"} Dec 04 17:58:46 crc kubenswrapper[4733]: I1204 17:58:46.344333 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d3a2277-a4db-4929-8fa2-67462cdc88fa" path="/var/lib/kubelet/pods/5d3a2277-a4db-4929-8fa2-67462cdc88fa/volumes" Dec 04 17:58:46 crc kubenswrapper[4733]: I1204 17:58:46.759422 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e605c62-64b4-4417-80bb-bc3387881f7a","Type":"ContainerStarted","Data":"9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.800162 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" event={"ID":"e659c44b-3009-4c5a-8a49-2d338d0d4c37","Type":"ContainerStarted","Data":"65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.800947 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.805282 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" event={"ID":"a35bc564-4823-4bb3-9a04-7e1bdcc56a17","Type":"ContainerStarted","Data":"46128b493a7ea8a8202efe06da8905c951a3b62403f1160fd1690a7f41d05ba3"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.805596 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.807958 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3d28f771-2fba-42d9-8fc0-1107d17db527","Type":"ContainerStarted","Data":"5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.808339 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.810829 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"47e99629-3bc7-4756-9b65-11ab755f49ad","Type":"ContainerStarted","Data":"29a65e283a9d82be93f99207a11c39d986879b38e8854120db8b1b0e54553298"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.811453 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.813706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"229fa5f0-e006-4c04-8476-23a7842d292b","Type":"ContainerStarted","Data":"c3b464a1016d052759179bfb0e7253393babdea9ae2efb95d1c7ade0f58ff289"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.813740 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"229fa5f0-e006-4c04-8476-23a7842d292b","Type":"ContainerStarted","Data":"11928a6e357a5e63f468c1daf6d2ee76e316437dd96f549ec172b0d953d18271"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.813941 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.815934 4733 generic.go:334] "Generic (PLEG): container finished" podID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerID="f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6" exitCode=0 Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.816076 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9cb46d6-3619-4035-8431-0b8552c5e690","Type":"ContainerDied","Data":"f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.821473 4733 generic.go:334] "Generic (PLEG): container finished" podID="5833da06-817f-4451-be68-9a9a94e5ef64" containerID="9ccae8be73d6c45906bab0a51e77d0e62fe11a561821293fc12bd3c1b094fd50" exitCode=0 Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.821536 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5833da06-817f-4451-be68-9a9a94e5ef64","Type":"ContainerDied","Data":"9ccae8be73d6c45906bab0a51e77d0e62fe11a561821293fc12bd3c1b094fd50"} Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.848568 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" podStartSLOduration=7.11114237 podStartE2EDuration="7.848526327s" podCreationTimestamp="2025-12-04 17:58:43 +0000 UTC" firstStartedPulling="2025-12-04 17:58:44.355017677 +0000 UTC m=+1186.310378723" lastFinishedPulling="2025-12-04 17:58:45.092401634 +0000 UTC m=+1187.047762680" observedRunningTime="2025-12-04 17:58:50.843963605 +0000 UTC m=+1192.799324711" watchObservedRunningTime="2025-12-04 17:58:50.848526327 +0000 UTC m=+1192.803887423" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.932026 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" podStartSLOduration=7.313922462 podStartE2EDuration="7.932004057s" podCreationTimestamp="2025-12-04 17:58:43 +0000 UTC" firstStartedPulling="2025-12-04 17:58:44.476064096 +0000 UTC m=+1186.431425142" lastFinishedPulling="2025-12-04 17:58:45.094145691 +0000 UTC m=+1187.049506737" observedRunningTime="2025-12-04 17:58:50.924475646 +0000 UTC m=+1192.879836702" watchObservedRunningTime="2025-12-04 17:58:50.932004057 +0000 UTC m=+1192.887365133" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.934780 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.570172398 podStartE2EDuration="7.93476465s" podCreationTimestamp="2025-12-04 17:58:43 +0000 UTC" firstStartedPulling="2025-12-04 17:58:44.838705549 +0000 UTC m=+1186.794066595" lastFinishedPulling="2025-12-04 17:58:49.203297791 +0000 UTC m=+1191.158658847" observedRunningTime="2025-12-04 17:58:50.899881052 +0000 UTC m=+1192.855242098" watchObservedRunningTime="2025-12-04 17:58:50.93476465 +0000 UTC m=+1192.890125726" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.949157 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.331690182 podStartE2EDuration="51.949137962s" podCreationTimestamp="2025-12-04 17:57:59 +0000 UTC" firstStartedPulling="2025-12-04 17:58:00.382641977 +0000 UTC m=+1142.338003023" lastFinishedPulling="2025-12-04 17:58:50.000089737 +0000 UTC m=+1191.955450803" observedRunningTime="2025-12-04 17:58:50.940327518 +0000 UTC m=+1192.895688604" watchObservedRunningTime="2025-12-04 17:58:50.949137962 +0000 UTC m=+1192.904499018" Dec 04 17:58:50 crc kubenswrapper[4733]: I1204 17:58:50.968825 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.437041555 podStartE2EDuration="53.968804825s" podCreationTimestamp="2025-12-04 17:57:57 +0000 UTC" firstStartedPulling="2025-12-04 17:57:58.295942704 +0000 UTC m=+1140.251303750" lastFinishedPulling="2025-12-04 17:58:49.827705974 +0000 UTC m=+1191.783067020" observedRunningTime="2025-12-04 17:58:50.966917195 +0000 UTC m=+1192.922278261" watchObservedRunningTime="2025-12-04 17:58:50.968804825 +0000 UTC m=+1192.924165881" Dec 04 17:58:51 crc kubenswrapper[4733]: I1204 17:58:51.846878 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9cb46d6-3619-4035-8431-0b8552c5e690","Type":"ContainerStarted","Data":"e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c"} Dec 04 17:58:51 crc kubenswrapper[4733]: I1204 17:58:51.859552 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5833da06-817f-4451-be68-9a9a94e5ef64","Type":"ContainerStarted","Data":"51718f0a98bc4f12648ae0c8e161095d9435571807a28e34ed23fb1ce1e33217"} Dec 04 17:58:51 crc kubenswrapper[4733]: I1204 17:58:51.863666 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f55e362-d59c-4269-92c3-d5ca014a2ef1","Type":"ContainerStarted","Data":"d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce"} Dec 04 17:58:51 crc kubenswrapper[4733]: I1204 17:58:51.897595 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.944221829 podStartE2EDuration="57.897565741s" podCreationTimestamp="2025-12-04 17:57:54 +0000 UTC" firstStartedPulling="2025-12-04 17:57:56.52895527 +0000 UTC m=+1138.484316316" lastFinishedPulling="2025-12-04 17:58:44.482299182 +0000 UTC m=+1186.437660228" observedRunningTime="2025-12-04 17:58:51.88701427 +0000 UTC m=+1193.842375366" watchObservedRunningTime="2025-12-04 17:58:51.897565741 +0000 UTC m=+1193.852926817" Dec 04 17:58:51 crc kubenswrapper[4733]: I1204 17:58:51.922172 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.308535764 podStartE2EDuration="56.922140883s" podCreationTimestamp="2025-12-04 17:57:55 +0000 UTC" firstStartedPulling="2025-12-04 17:57:58.141890607 +0000 UTC m=+1140.097251653" lastFinishedPulling="2025-12-04 17:58:44.755495726 +0000 UTC m=+1186.710856772" observedRunningTime="2025-12-04 17:58:51.908161722 +0000 UTC m=+1193.863522778" watchObservedRunningTime="2025-12-04 17:58:51.922140883 +0000 UTC m=+1193.877501969" Dec 04 17:58:55 crc kubenswrapper[4733]: I1204 17:58:55.903947 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 04 17:58:55 crc kubenswrapper[4733]: I1204 17:58:55.904339 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 04 17:58:55 crc kubenswrapper[4733]: I1204 17:58:55.979284 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.054294 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.326767 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-df92-account-create-update-p899x"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.327955 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.331886 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.343634 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-df92-account-create-update-p899x"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.360632 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj4c7\" (UniqueName: \"kubernetes.io/projected/16f02d93-bf6f-4123-8781-2238ffbac212-kube-api-access-rj4c7\") pod \"keystone-df92-account-create-update-p899x\" (UID: \"16f02d93-bf6f-4123-8781-2238ffbac212\") " pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.360684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f02d93-bf6f-4123-8781-2238ffbac212-operator-scripts\") pod \"keystone-df92-account-create-update-p899x\" (UID: \"16f02d93-bf6f-4123-8781-2238ffbac212\") " pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.360719 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.361004 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.383476 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-b92dw"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.384503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-b92dw" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.392108 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-b92dw"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.463655 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj4c7\" (UniqueName: \"kubernetes.io/projected/16f02d93-bf6f-4123-8781-2238ffbac212-kube-api-access-rj4c7\") pod \"keystone-df92-account-create-update-p899x\" (UID: \"16f02d93-bf6f-4123-8781-2238ffbac212\") " pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.463742 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f02d93-bf6f-4123-8781-2238ffbac212-operator-scripts\") pod \"keystone-df92-account-create-update-p899x\" (UID: \"16f02d93-bf6f-4123-8781-2238ffbac212\") " pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.463854 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnnrn\" (UniqueName: \"kubernetes.io/projected/a6b87ca2-4352-4387-8e5d-6070a278c63d-kube-api-access-rnnrn\") pod \"keystone-db-create-b92dw\" (UID: \"a6b87ca2-4352-4387-8e5d-6070a278c63d\") " pod="openstack/keystone-db-create-b92dw" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.464216 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6b87ca2-4352-4387-8e5d-6070a278c63d-operator-scripts\") pod \"keystone-db-create-b92dw\" (UID: \"a6b87ca2-4352-4387-8e5d-6070a278c63d\") " pod="openstack/keystone-db-create-b92dw" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.464664 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f02d93-bf6f-4123-8781-2238ffbac212-operator-scripts\") pod \"keystone-df92-account-create-update-p899x\" (UID: \"16f02d93-bf6f-4123-8781-2238ffbac212\") " pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.481401 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj4c7\" (UniqueName: \"kubernetes.io/projected/16f02d93-bf6f-4123-8781-2238ffbac212-kube-api-access-rj4c7\") pod \"keystone-df92-account-create-update-p899x\" (UID: \"16f02d93-bf6f-4123-8781-2238ffbac212\") " pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.566200 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6b87ca2-4352-4387-8e5d-6070a278c63d-operator-scripts\") pod \"keystone-db-create-b92dw\" (UID: \"a6b87ca2-4352-4387-8e5d-6070a278c63d\") " pod="openstack/keystone-db-create-b92dw" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.566302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnnrn\" (UniqueName: \"kubernetes.io/projected/a6b87ca2-4352-4387-8e5d-6070a278c63d-kube-api-access-rnnrn\") pod \"keystone-db-create-b92dw\" (UID: \"a6b87ca2-4352-4387-8e5d-6070a278c63d\") " pod="openstack/keystone-db-create-b92dw" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.566653 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-zn7w9"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.567467 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6b87ca2-4352-4387-8e5d-6070a278c63d-operator-scripts\") pod \"keystone-db-create-b92dw\" (UID: \"a6b87ca2-4352-4387-8e5d-6070a278c63d\") " pod="openstack/keystone-db-create-b92dw" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.567870 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-zn7w9" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.575101 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-zn7w9"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.593569 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnnrn\" (UniqueName: \"kubernetes.io/projected/a6b87ca2-4352-4387-8e5d-6070a278c63d-kube-api-access-rnnrn\") pod \"keystone-db-create-b92dw\" (UID: \"a6b87ca2-4352-4387-8e5d-6070a278c63d\") " pod="openstack/keystone-db-create-b92dw" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.650466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.672814 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a32f53d-9569-405a-9060-80e298dfc6ce-operator-scripts\") pod \"placement-db-create-zn7w9\" (UID: \"6a32f53d-9569-405a-9060-80e298dfc6ce\") " pod="openstack/placement-db-create-zn7w9" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.673223 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x85h8\" (UniqueName: \"kubernetes.io/projected/6a32f53d-9569-405a-9060-80e298dfc6ce-kube-api-access-x85h8\") pod \"placement-db-create-zn7w9\" (UID: \"6a32f53d-9569-405a-9060-80e298dfc6ce\") " pod="openstack/placement-db-create-zn7w9" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.675614 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ee67-account-create-update-56lqg"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.676855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.680705 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.683009 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.687906 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ee67-account-create-update-56lqg"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.708251 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-b92dw" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.774446 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a32f53d-9569-405a-9060-80e298dfc6ce-operator-scripts\") pod \"placement-db-create-zn7w9\" (UID: \"6a32f53d-9569-405a-9060-80e298dfc6ce\") " pod="openstack/placement-db-create-zn7w9" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.774508 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x85h8\" (UniqueName: \"kubernetes.io/projected/6a32f53d-9569-405a-9060-80e298dfc6ce-kube-api-access-x85h8\") pod \"placement-db-create-zn7w9\" (UID: \"6a32f53d-9569-405a-9060-80e298dfc6ce\") " pod="openstack/placement-db-create-zn7w9" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.774639 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b41b1c63-78a9-416c-9358-9a436fae72da-operator-scripts\") pod \"placement-ee67-account-create-update-56lqg\" (UID: \"b41b1c63-78a9-416c-9358-9a436fae72da\") " pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.774747 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thk49\" (UniqueName: \"kubernetes.io/projected/b41b1c63-78a9-416c-9358-9a436fae72da-kube-api-access-thk49\") pod \"placement-ee67-account-create-update-56lqg\" (UID: \"b41b1c63-78a9-416c-9358-9a436fae72da\") " pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.776087 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a32f53d-9569-405a-9060-80e298dfc6ce-operator-scripts\") pod \"placement-db-create-zn7w9\" (UID: \"6a32f53d-9569-405a-9060-80e298dfc6ce\") " pod="openstack/placement-db-create-zn7w9" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.804159 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x85h8\" (UniqueName: \"kubernetes.io/projected/6a32f53d-9569-405a-9060-80e298dfc6ce-kube-api-access-x85h8\") pod \"placement-db-create-zn7w9\" (UID: \"6a32f53d-9569-405a-9060-80e298dfc6ce\") " pod="openstack/placement-db-create-zn7w9" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.877593 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b41b1c63-78a9-416c-9358-9a436fae72da-operator-scripts\") pod \"placement-ee67-account-create-update-56lqg\" (UID: \"b41b1c63-78a9-416c-9358-9a436fae72da\") " pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.877930 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thk49\" (UniqueName: \"kubernetes.io/projected/b41b1c63-78a9-416c-9358-9a436fae72da-kube-api-access-thk49\") pod \"placement-ee67-account-create-update-56lqg\" (UID: \"b41b1c63-78a9-416c-9358-9a436fae72da\") " pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.879153 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b41b1c63-78a9-416c-9358-9a436fae72da-operator-scripts\") pod \"placement-ee67-account-create-update-56lqg\" (UID: \"b41b1c63-78a9-416c-9358-9a436fae72da\") " pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.886644 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-zn7w9" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.898621 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-t67cq"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.899625 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-t67cq" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.900183 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thk49\" (UniqueName: \"kubernetes.io/projected/b41b1c63-78a9-416c-9358-9a436fae72da-kube-api-access-thk49\") pod \"placement-ee67-account-create-update-56lqg\" (UID: \"b41b1c63-78a9-416c-9358-9a436fae72da\") " pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.935763 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-t67cq"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.970506 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-df92-account-create-update-p899x"] Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.980264 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b35785d-5a00-4554-b0ac-5585ca1c4816-operator-scripts\") pod \"glance-db-create-t67cq\" (UID: \"1b35785d-5a00-4554-b0ac-5585ca1c4816\") " pod="openstack/glance-db-create-t67cq" Dec 04 17:58:57 crc kubenswrapper[4733]: I1204 17:58:57.980311 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j28pk\" (UniqueName: \"kubernetes.io/projected/1b35785d-5a00-4554-b0ac-5585ca1c4816-kube-api-access-j28pk\") pod \"glance-db-create-t67cq\" (UID: \"1b35785d-5a00-4554-b0ac-5585ca1c4816\") " pod="openstack/glance-db-create-t67cq" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.001379 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-776f-account-create-update-cfcmh"] Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.002268 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.004981 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.012705 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-776f-account-create-update-cfcmh"] Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.044416 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.081473 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-operator-scripts\") pod \"glance-776f-account-create-update-cfcmh\" (UID: \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\") " pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.081526 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b35785d-5a00-4554-b0ac-5585ca1c4816-operator-scripts\") pod \"glance-db-create-t67cq\" (UID: \"1b35785d-5a00-4554-b0ac-5585ca1c4816\") " pod="openstack/glance-db-create-t67cq" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.081565 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j28pk\" (UniqueName: \"kubernetes.io/projected/1b35785d-5a00-4554-b0ac-5585ca1c4816-kube-api-access-j28pk\") pod \"glance-db-create-t67cq\" (UID: \"1b35785d-5a00-4554-b0ac-5585ca1c4816\") " pod="openstack/glance-db-create-t67cq" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.081644 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9vxr\" (UniqueName: \"kubernetes.io/projected/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-kube-api-access-q9vxr\") pod \"glance-776f-account-create-update-cfcmh\" (UID: \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\") " pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.083320 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b35785d-5a00-4554-b0ac-5585ca1c4816-operator-scripts\") pod \"glance-db-create-t67cq\" (UID: \"1b35785d-5a00-4554-b0ac-5585ca1c4816\") " pod="openstack/glance-db-create-t67cq" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.098777 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j28pk\" (UniqueName: \"kubernetes.io/projected/1b35785d-5a00-4554-b0ac-5585ca1c4816-kube-api-access-j28pk\") pod \"glance-db-create-t67cq\" (UID: \"1b35785d-5a00-4554-b0ac-5585ca1c4816\") " pod="openstack/glance-db-create-t67cq" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.103759 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.132113 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.184062 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-operator-scripts\") pod \"glance-776f-account-create-update-cfcmh\" (UID: \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\") " pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.184768 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9vxr\" (UniqueName: \"kubernetes.io/projected/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-kube-api-access-q9vxr\") pod \"glance-776f-account-create-update-cfcmh\" (UID: \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\") " pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.186331 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-operator-scripts\") pod \"glance-776f-account-create-update-cfcmh\" (UID: \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\") " pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.203963 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9vxr\" (UniqueName: \"kubernetes.io/projected/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-kube-api-access-q9vxr\") pod \"glance-776f-account-create-update-cfcmh\" (UID: \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\") " pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.221722 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-t67cq" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.258511 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-b92dw"] Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.322898 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.339900 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-zn7w9"] Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.551404 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ee67-account-create-update-56lqg"] Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.567134 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.688519 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-t67cq"] Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.831962 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.913406 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-776f-account-create-update-cfcmh"] Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.957089 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.957138 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.969555 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ee67-account-create-update-56lqg" event={"ID":"b41b1c63-78a9-416c-9358-9a436fae72da","Type":"ContainerStarted","Data":"c7c213db13cc2a1031c50a113eaf1ccb9e27d3fc38121005725b7eea14e62d2e"} Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.976073 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-776f-account-create-update-cfcmh" event={"ID":"62b8dc93-1d10-4355-9d7f-ce423a3c11bc","Type":"ContainerStarted","Data":"1e66593d347930c022495875cda2be4d8793bcd50e28eb56a3c6fb2a8ea390fc"} Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.984323 4733 generic.go:334] "Generic (PLEG): container finished" podID="16f02d93-bf6f-4123-8781-2238ffbac212" containerID="7c64cb40e986cfa71acbcbd9c5202683bf86fb0a870830d77be7d32b48479000" exitCode=0 Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.984479 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-df92-account-create-update-p899x" event={"ID":"16f02d93-bf6f-4123-8781-2238ffbac212","Type":"ContainerDied","Data":"7c64cb40e986cfa71acbcbd9c5202683bf86fb0a870830d77be7d32b48479000"} Dec 04 17:58:58 crc kubenswrapper[4733]: I1204 17:58:58.984544 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-df92-account-create-update-p899x" event={"ID":"16f02d93-bf6f-4123-8781-2238ffbac212","Type":"ContainerStarted","Data":"a68a756253d40724bb7b21be1ea6ed2dc25dda6abe2c7efcedc98bff42a00dfc"} Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:58.996757 4733 generic.go:334] "Generic (PLEG): container finished" podID="a6b87ca2-4352-4387-8e5d-6070a278c63d" containerID="4a1a626a6d350a4c0d0cb45c40052b41650203413c0c302e9f5fe51aba7cd968" exitCode=0 Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:58.996977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-b92dw" event={"ID":"a6b87ca2-4352-4387-8e5d-6070a278c63d","Type":"ContainerDied","Data":"4a1a626a6d350a4c0d0cb45c40052b41650203413c0c302e9f5fe51aba7cd968"} Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:58.997008 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-b92dw" event={"ID":"a6b87ca2-4352-4387-8e5d-6070a278c63d","Type":"ContainerStarted","Data":"9734b8cfd7d95fe2849a934e8fb9d330453373d9599c817fff5ab441de4c3ba9"} Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.005567 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-zn7w9" event={"ID":"6a32f53d-9569-405a-9060-80e298dfc6ce","Type":"ContainerStarted","Data":"4c0f39dd57ee5768e8cb2fc129fedea35799ad4350c78f9e6bf0de97e2423216"} Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.005621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-zn7w9" event={"ID":"6a32f53d-9569-405a-9060-80e298dfc6ce","Type":"ContainerStarted","Data":"116cd766a36ed2dd12e15239427dbb6fcf8e1a3a89818c9b2d0406e54b2e1ce1"} Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.030827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-t67cq" event={"ID":"1b35785d-5a00-4554-b0ac-5585ca1c4816","Type":"ContainerStarted","Data":"867c3a2cd00deeb9bc457acc31e40444077d73f7cd1cb93e3ea5976bc61e50e5"} Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.037808 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-dntx9"] Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.038026 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" podUID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" containerName="dnsmasq-dns" containerID="cri-o://65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25" gracePeriod=10 Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.068181 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-zn7w9" podStartSLOduration=2.068162292 podStartE2EDuration="2.068162292s" podCreationTimestamp="2025-12-04 17:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:58:59.050219075 +0000 UTC m=+1201.005580131" watchObservedRunningTime="2025-12-04 17:58:59.068162292 +0000 UTC m=+1201.023523338" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.079342 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-t67cq" podStartSLOduration=2.079323739 podStartE2EDuration="2.079323739s" podCreationTimestamp="2025-12-04 17:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:58:59.064423573 +0000 UTC m=+1201.019784639" watchObservedRunningTime="2025-12-04 17:58:59.079323739 +0000 UTC m=+1201.034684785" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.442691 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.612429 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-ovsdbserver-sb\") pod \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.612519 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-dns-svc\") pod \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.612541 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-config\") pod \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.612613 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64562\" (UniqueName: \"kubernetes.io/projected/e659c44b-3009-4c5a-8a49-2d338d0d4c37-kube-api-access-64562\") pod \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\" (UID: \"e659c44b-3009-4c5a-8a49-2d338d0d4c37\") " Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.623077 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e659c44b-3009-4c5a-8a49-2d338d0d4c37-kube-api-access-64562" (OuterVolumeSpecName: "kube-api-access-64562") pod "e659c44b-3009-4c5a-8a49-2d338d0d4c37" (UID: "e659c44b-3009-4c5a-8a49-2d338d0d4c37"). InnerVolumeSpecName "kube-api-access-64562". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.664593 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e659c44b-3009-4c5a-8a49-2d338d0d4c37" (UID: "e659c44b-3009-4c5a-8a49-2d338d0d4c37"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.671419 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-config" (OuterVolumeSpecName: "config") pod "e659c44b-3009-4c5a-8a49-2d338d0d4c37" (UID: "e659c44b-3009-4c5a-8a49-2d338d0d4c37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.695550 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e659c44b-3009-4c5a-8a49-2d338d0d4c37" (UID: "e659c44b-3009-4c5a-8a49-2d338d0d4c37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.714241 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.714287 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.714303 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e659c44b-3009-4c5a-8a49-2d338d0d4c37-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.714318 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64562\" (UniqueName: \"kubernetes.io/projected/e659c44b-3009-4c5a-8a49-2d338d0d4c37-kube-api-access-64562\") on node \"crc\" DevicePath \"\"" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.818041 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.863757 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-9b9k2"] Dec 04 17:58:59 crc kubenswrapper[4733]: E1204 17:58:59.864431 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" containerName="init" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.864444 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" containerName="init" Dec 04 17:58:59 crc kubenswrapper[4733]: E1204 17:58:59.864462 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" containerName="dnsmasq-dns" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.864469 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" containerName="dnsmasq-dns" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.864614 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" containerName="dnsmasq-dns" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.866054 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:58:59 crc kubenswrapper[4733]: I1204 17:58:59.878497 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-9b9k2"] Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.019138 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nptzp\" (UniqueName: \"kubernetes.io/projected/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-kube-api-access-nptzp\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.019226 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.019264 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.019294 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.019327 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-config\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.039199 4733 generic.go:334] "Generic (PLEG): container finished" podID="b41b1c63-78a9-416c-9358-9a436fae72da" containerID="da70a66a6d6be97c5ab3e3bb531110131ae53dcb8be21760ae2043a41665da7a" exitCode=0 Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.039257 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ee67-account-create-update-56lqg" event={"ID":"b41b1c63-78a9-416c-9358-9a436fae72da","Type":"ContainerDied","Data":"da70a66a6d6be97c5ab3e3bb531110131ae53dcb8be21760ae2043a41665da7a"} Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.042390 4733 generic.go:334] "Generic (PLEG): container finished" podID="62b8dc93-1d10-4355-9d7f-ce423a3c11bc" containerID="04e97a72c5dc8120564a0e524dfcc8fc7aae822e190797b30bde855daa1981d9" exitCode=0 Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.042430 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-776f-account-create-update-cfcmh" event={"ID":"62b8dc93-1d10-4355-9d7f-ce423a3c11bc","Type":"ContainerDied","Data":"04e97a72c5dc8120564a0e524dfcc8fc7aae822e190797b30bde855daa1981d9"} Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.045297 4733 generic.go:334] "Generic (PLEG): container finished" podID="6a32f53d-9569-405a-9060-80e298dfc6ce" containerID="4c0f39dd57ee5768e8cb2fc129fedea35799ad4350c78f9e6bf0de97e2423216" exitCode=0 Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.045336 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-zn7w9" event={"ID":"6a32f53d-9569-405a-9060-80e298dfc6ce","Type":"ContainerDied","Data":"4c0f39dd57ee5768e8cb2fc129fedea35799ad4350c78f9e6bf0de97e2423216"} Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.046875 4733 generic.go:334] "Generic (PLEG): container finished" podID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" containerID="65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25" exitCode=0 Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.046918 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" event={"ID":"e659c44b-3009-4c5a-8a49-2d338d0d4c37","Type":"ContainerDied","Data":"65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25"} Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.046934 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" event={"ID":"e659c44b-3009-4c5a-8a49-2d338d0d4c37","Type":"ContainerDied","Data":"c595dd82e504a002be34ad6ca1822cafd965d7b192a8a2c7d143b24e175f0dab"} Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.046949 4733 scope.go:117] "RemoveContainer" containerID="65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.047032 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb7d9565c-dntx9" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.050754 4733 generic.go:334] "Generic (PLEG): container finished" podID="1b35785d-5a00-4554-b0ac-5585ca1c4816" containerID="e2276e988658554a414066aa0968e027716dad08111b838f90813477030b0565" exitCode=0 Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.050980 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-t67cq" event={"ID":"1b35785d-5a00-4554-b0ac-5585ca1c4816","Type":"ContainerDied","Data":"e2276e988658554a414066aa0968e027716dad08111b838f90813477030b0565"} Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.095468 4733 scope.go:117] "RemoveContainer" containerID="8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.122106 4733 scope.go:117] "RemoveContainer" containerID="65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.122779 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nptzp\" (UniqueName: \"kubernetes.io/projected/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-kube-api-access-nptzp\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.122890 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.122943 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.122985 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-config\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.123005 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.123899 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.123903 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.124586 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-dns-svc\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.124727 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-config\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: E1204 17:59:00.138412 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25\": container with ID starting with 65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25 not found: ID does not exist" containerID="65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.138455 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25"} err="failed to get container status \"65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25\": rpc error: code = NotFound desc = could not find container \"65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25\": container with ID starting with 65088ff3801acec0d1706e43beb2cfc6d1987058b9b054694b88d329cbf0ee25 not found: ID does not exist" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.138481 4733 scope.go:117] "RemoveContainer" containerID="8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb" Dec 04 17:59:00 crc kubenswrapper[4733]: E1204 17:59:00.142973 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb\": container with ID starting with 8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb not found: ID does not exist" containerID="8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.143014 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb"} err="failed to get container status \"8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb\": rpc error: code = NotFound desc = could not find container \"8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb\": container with ID starting with 8564f78dc459951e590ebae74a71862ca1faf7cb6cf196ae8203f85af1831aeb not found: ID does not exist" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.146630 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nptzp\" (UniqueName: \"kubernetes.io/projected/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-kube-api-access-nptzp\") pod \"dnsmasq-dns-59d5fbdd8c-9b9k2\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.149360 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-dntx9"] Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.177176 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb7d9565c-dntx9"] Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.187747 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.368417 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e659c44b-3009-4c5a-8a49-2d338d0d4c37" path="/var/lib/kubelet/pods/e659c44b-3009-4c5a-8a49-2d338d0d4c37/volumes" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.411215 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.529901 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj4c7\" (UniqueName: \"kubernetes.io/projected/16f02d93-bf6f-4123-8781-2238ffbac212-kube-api-access-rj4c7\") pod \"16f02d93-bf6f-4123-8781-2238ffbac212\" (UID: \"16f02d93-bf6f-4123-8781-2238ffbac212\") " Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.530164 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f02d93-bf6f-4123-8781-2238ffbac212-operator-scripts\") pod \"16f02d93-bf6f-4123-8781-2238ffbac212\" (UID: \"16f02d93-bf6f-4123-8781-2238ffbac212\") " Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.532922 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f02d93-bf6f-4123-8781-2238ffbac212-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16f02d93-bf6f-4123-8781-2238ffbac212" (UID: "16f02d93-bf6f-4123-8781-2238ffbac212"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.537312 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16f02d93-bf6f-4123-8781-2238ffbac212-kube-api-access-rj4c7" (OuterVolumeSpecName: "kube-api-access-rj4c7") pod "16f02d93-bf6f-4123-8781-2238ffbac212" (UID: "16f02d93-bf6f-4123-8781-2238ffbac212"). InnerVolumeSpecName "kube-api-access-rj4c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.611596 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-b92dw" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.634682 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16f02d93-bf6f-4123-8781-2238ffbac212-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.634727 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj4c7\" (UniqueName: \"kubernetes.io/projected/16f02d93-bf6f-4123-8781-2238ffbac212-kube-api-access-rj4c7\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.694727 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-9b9k2"] Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.735677 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnnrn\" (UniqueName: \"kubernetes.io/projected/a6b87ca2-4352-4387-8e5d-6070a278c63d-kube-api-access-rnnrn\") pod \"a6b87ca2-4352-4387-8e5d-6070a278c63d\" (UID: \"a6b87ca2-4352-4387-8e5d-6070a278c63d\") " Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.736105 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6b87ca2-4352-4387-8e5d-6070a278c63d-operator-scripts\") pod \"a6b87ca2-4352-4387-8e5d-6070a278c63d\" (UID: \"a6b87ca2-4352-4387-8e5d-6070a278c63d\") " Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.736875 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6b87ca2-4352-4387-8e5d-6070a278c63d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6b87ca2-4352-4387-8e5d-6070a278c63d" (UID: "a6b87ca2-4352-4387-8e5d-6070a278c63d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.740026 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6b87ca2-4352-4387-8e5d-6070a278c63d-kube-api-access-rnnrn" (OuterVolumeSpecName: "kube-api-access-rnnrn") pod "a6b87ca2-4352-4387-8e5d-6070a278c63d" (UID: "a6b87ca2-4352-4387-8e5d-6070a278c63d"). InnerVolumeSpecName "kube-api-access-rnnrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.837696 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6b87ca2-4352-4387-8e5d-6070a278c63d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:00 crc kubenswrapper[4733]: I1204 17:59:00.837734 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnnrn\" (UniqueName: \"kubernetes.io/projected/a6b87ca2-4352-4387-8e5d-6070a278c63d-kube-api-access-rnnrn\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.041391 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.042734 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6b87ca2-4352-4387-8e5d-6070a278c63d" containerName="mariadb-database-create" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.042767 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6b87ca2-4352-4387-8e5d-6070a278c63d" containerName="mariadb-database-create" Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.042873 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f02d93-bf6f-4123-8781-2238ffbac212" containerName="mariadb-account-create-update" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.042891 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f02d93-bf6f-4123-8781-2238ffbac212" containerName="mariadb-account-create-update" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.043525 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="16f02d93-bf6f-4123-8781-2238ffbac212" containerName="mariadb-account-create-update" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.043586 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6b87ca2-4352-4387-8e5d-6070a278c63d" containerName="mariadb-database-create" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.059281 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.065604 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.105682 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.105768 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-x84j5" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.105887 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.106074 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.116003 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-df92-account-create-update-p899x" event={"ID":"16f02d93-bf6f-4123-8781-2238ffbac212","Type":"ContainerDied","Data":"a68a756253d40724bb7b21be1ea6ed2dc25dda6abe2c7efcedc98bff42a00dfc"} Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.116063 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a68a756253d40724bb7b21be1ea6ed2dc25dda6abe2c7efcedc98bff42a00dfc" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.116130 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-df92-account-create-update-p899x" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.133031 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-b92dw" event={"ID":"a6b87ca2-4352-4387-8e5d-6070a278c63d","Type":"ContainerDied","Data":"9734b8cfd7d95fe2849a934e8fb9d330453373d9599c817fff5ab441de4c3ba9"} Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.133075 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9734b8cfd7d95fe2849a934e8fb9d330453373d9599c817fff5ab441de4c3ba9" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.133196 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-b92dw" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.142605 4733 generic.go:334] "Generic (PLEG): container finished" podID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" containerID="250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12" exitCode=0 Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.142688 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" event={"ID":"51afaec5-8ac0-47ab-a960-63bb01cf4fc1","Type":"ContainerDied","Data":"250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12"} Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.142712 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" event={"ID":"51afaec5-8ac0-47ab-a960-63bb01cf4fc1","Type":"ContainerStarted","Data":"0ff7a5bd86ca0753d79b844ffabd02f884b527ac493b22c4142ef691f3bf1072"} Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.250639 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.250738 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpxbm\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-kube-api-access-qpxbm\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.250810 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-lock\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.250837 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.250864 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-cache\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.352346 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-cache\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.352428 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.352475 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpxbm\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-kube-api-access-qpxbm\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.352513 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-lock\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.352536 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.352925 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.352948 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.352992 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift podName:75da96fa-8079-442d-8993-de5658a23f91 nodeName:}" failed. No retries permitted until 2025-12-04 17:59:01.852976714 +0000 UTC m=+1203.808337760 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift") pod "swift-storage-0" (UID: "75da96fa-8079-442d-8993-de5658a23f91") : configmap "swift-ring-files" not found Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.352934 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.353864 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-lock\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.353995 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-cache\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.374688 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.375739 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpxbm\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-kube-api-access-qpxbm\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.475982 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.581729 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9t2z2"] Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.582122 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b41b1c63-78a9-416c-9358-9a436fae72da" containerName="mariadb-account-create-update" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.582140 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b41b1c63-78a9-416c-9358-9a436fae72da" containerName="mariadb-account-create-update" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.582370 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b41b1c63-78a9-416c-9358-9a436fae72da" containerName="mariadb-account-create-update" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.583107 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.585680 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.586081 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.586186 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.596871 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9t2z2"] Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664210 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thk49\" (UniqueName: \"kubernetes.io/projected/b41b1c63-78a9-416c-9358-9a436fae72da-kube-api-access-thk49\") pod \"b41b1c63-78a9-416c-9358-9a436fae72da\" (UID: \"b41b1c63-78a9-416c-9358-9a436fae72da\") " Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664287 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b41b1c63-78a9-416c-9358-9a436fae72da-operator-scripts\") pod \"b41b1c63-78a9-416c-9358-9a436fae72da\" (UID: \"b41b1c63-78a9-416c-9358-9a436fae72da\") " Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664380 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-swiftconf\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664446 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-ring-data-devices\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664472 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqg8x\" (UniqueName: \"kubernetes.io/projected/3590b958-1ae5-46f6-9664-d8efdec482b2-kube-api-access-bqg8x\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664495 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-dispersionconf\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664526 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-combined-ca-bundle\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664542 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-scripts\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.664567 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3590b958-1ae5-46f6-9664-d8efdec482b2-etc-swift\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.665235 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b41b1c63-78a9-416c-9358-9a436fae72da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b41b1c63-78a9-416c-9358-9a436fae72da" (UID: "b41b1c63-78a9-416c-9358-9a436fae72da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.668365 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b41b1c63-78a9-416c-9358-9a436fae72da-kube-api-access-thk49" (OuterVolumeSpecName: "kube-api-access-thk49") pod "b41b1c63-78a9-416c-9358-9a436fae72da" (UID: "b41b1c63-78a9-416c-9358-9a436fae72da"). InnerVolumeSpecName "kube-api-access-thk49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.701060 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.718575 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-zn7w9" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.724760 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-t67cq" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.765736 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-swiftconf\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.765884 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-ring-data-devices\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.765924 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqg8x\" (UniqueName: \"kubernetes.io/projected/3590b958-1ae5-46f6-9664-d8efdec482b2-kube-api-access-bqg8x\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.766320 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-dispersionconf\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.766404 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-combined-ca-bundle\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.766429 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-scripts\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.766491 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3590b958-1ae5-46f6-9664-d8efdec482b2-etc-swift\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.766511 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-ring-data-devices\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.767507 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-scripts\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.767573 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thk49\" (UniqueName: \"kubernetes.io/projected/b41b1c63-78a9-416c-9358-9a436fae72da-kube-api-access-thk49\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.767593 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b41b1c63-78a9-416c-9358-9a436fae72da-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.768618 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3590b958-1ae5-46f6-9664-d8efdec482b2-etc-swift\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.769022 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-dispersionconf\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.770014 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-swiftconf\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.770323 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-combined-ca-bundle\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.780195 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqg8x\" (UniqueName: \"kubernetes.io/projected/3590b958-1ae5-46f6-9664-d8efdec482b2-kube-api-access-bqg8x\") pod \"swift-ring-rebalance-9t2z2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.868118 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j28pk\" (UniqueName: \"kubernetes.io/projected/1b35785d-5a00-4554-b0ac-5585ca1c4816-kube-api-access-j28pk\") pod \"1b35785d-5a00-4554-b0ac-5585ca1c4816\" (UID: \"1b35785d-5a00-4554-b0ac-5585ca1c4816\") " Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.868179 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a32f53d-9569-405a-9060-80e298dfc6ce-operator-scripts\") pod \"6a32f53d-9569-405a-9060-80e298dfc6ce\" (UID: \"6a32f53d-9569-405a-9060-80e298dfc6ce\") " Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.868268 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-operator-scripts\") pod \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\" (UID: \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\") " Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.868301 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b35785d-5a00-4554-b0ac-5585ca1c4816-operator-scripts\") pod \"1b35785d-5a00-4554-b0ac-5585ca1c4816\" (UID: \"1b35785d-5a00-4554-b0ac-5585ca1c4816\") " Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.868345 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9vxr\" (UniqueName: \"kubernetes.io/projected/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-kube-api-access-q9vxr\") pod \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\" (UID: \"62b8dc93-1d10-4355-9d7f-ce423a3c11bc\") " Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.868375 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x85h8\" (UniqueName: \"kubernetes.io/projected/6a32f53d-9569-405a-9060-80e298dfc6ce-kube-api-access-x85h8\") pod \"6a32f53d-9569-405a-9060-80e298dfc6ce\" (UID: \"6a32f53d-9569-405a-9060-80e298dfc6ce\") " Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.868592 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.868778 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.868820 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 17:59:01 crc kubenswrapper[4733]: E1204 17:59:01.868873 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift podName:75da96fa-8079-442d-8993-de5658a23f91 nodeName:}" failed. No retries permitted until 2025-12-04 17:59:02.868854041 +0000 UTC m=+1204.824215097 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift") pod "swift-storage-0" (UID: "75da96fa-8079-442d-8993-de5658a23f91") : configmap "swift-ring-files" not found Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.868871 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a32f53d-9569-405a-9060-80e298dfc6ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6a32f53d-9569-405a-9060-80e298dfc6ce" (UID: "6a32f53d-9569-405a-9060-80e298dfc6ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.869065 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b35785d-5a00-4554-b0ac-5585ca1c4816-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1b35785d-5a00-4554-b0ac-5585ca1c4816" (UID: "1b35785d-5a00-4554-b0ac-5585ca1c4816"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.869409 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "62b8dc93-1d10-4355-9d7f-ce423a3c11bc" (UID: "62b8dc93-1d10-4355-9d7f-ce423a3c11bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.873258 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-kube-api-access-q9vxr" (OuterVolumeSpecName: "kube-api-access-q9vxr") pod "62b8dc93-1d10-4355-9d7f-ce423a3c11bc" (UID: "62b8dc93-1d10-4355-9d7f-ce423a3c11bc"). InnerVolumeSpecName "kube-api-access-q9vxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.873874 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b35785d-5a00-4554-b0ac-5585ca1c4816-kube-api-access-j28pk" (OuterVolumeSpecName: "kube-api-access-j28pk") pod "1b35785d-5a00-4554-b0ac-5585ca1c4816" (UID: "1b35785d-5a00-4554-b0ac-5585ca1c4816"). InnerVolumeSpecName "kube-api-access-j28pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.876710 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a32f53d-9569-405a-9060-80e298dfc6ce-kube-api-access-x85h8" (OuterVolumeSpecName: "kube-api-access-x85h8") pod "6a32f53d-9569-405a-9060-80e298dfc6ce" (UID: "6a32f53d-9569-405a-9060-80e298dfc6ce"). InnerVolumeSpecName "kube-api-access-x85h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.969910 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9vxr\" (UniqueName: \"kubernetes.io/projected/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-kube-api-access-q9vxr\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.969953 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x85h8\" (UniqueName: \"kubernetes.io/projected/6a32f53d-9569-405a-9060-80e298dfc6ce-kube-api-access-x85h8\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.969962 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j28pk\" (UniqueName: \"kubernetes.io/projected/1b35785d-5a00-4554-b0ac-5585ca1c4816-kube-api-access-j28pk\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.969972 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a32f53d-9569-405a-9060-80e298dfc6ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.969981 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b8dc93-1d10-4355-9d7f-ce423a3c11bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:01 crc kubenswrapper[4733]: I1204 17:59:01.969989 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b35785d-5a00-4554-b0ac-5585ca1c4816-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.015831 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.160726 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-t67cq" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.160718 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-t67cq" event={"ID":"1b35785d-5a00-4554-b0ac-5585ca1c4816","Type":"ContainerDied","Data":"867c3a2cd00deeb9bc457acc31e40444077d73f7cd1cb93e3ea5976bc61e50e5"} Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.161200 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="867c3a2cd00deeb9bc457acc31e40444077d73f7cd1cb93e3ea5976bc61e50e5" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.162966 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ee67-account-create-update-56lqg" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.163032 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ee67-account-create-update-56lqg" event={"ID":"b41b1c63-78a9-416c-9358-9a436fae72da","Type":"ContainerDied","Data":"c7c213db13cc2a1031c50a113eaf1ccb9e27d3fc38121005725b7eea14e62d2e"} Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.163055 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7c213db13cc2a1031c50a113eaf1ccb9e27d3fc38121005725b7eea14e62d2e" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.166252 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-776f-account-create-update-cfcmh" event={"ID":"62b8dc93-1d10-4355-9d7f-ce423a3c11bc","Type":"ContainerDied","Data":"1e66593d347930c022495875cda2be4d8793bcd50e28eb56a3c6fb2a8ea390fc"} Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.166297 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e66593d347930c022495875cda2be4d8793bcd50e28eb56a3c6fb2a8ea390fc" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.166547 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-776f-account-create-update-cfcmh" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.170408 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" event={"ID":"51afaec5-8ac0-47ab-a960-63bb01cf4fc1","Type":"ContainerStarted","Data":"42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435"} Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.170548 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.174533 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-zn7w9" event={"ID":"6a32f53d-9569-405a-9060-80e298dfc6ce","Type":"ContainerDied","Data":"116cd766a36ed2dd12e15239427dbb6fcf8e1a3a89818c9b2d0406e54b2e1ce1"} Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.174599 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="116cd766a36ed2dd12e15239427dbb6fcf8e1a3a89818c9b2d0406e54b2e1ce1" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.174681 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-zn7w9" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.208386 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" podStartSLOduration=3.208368748 podStartE2EDuration="3.208368748s" podCreationTimestamp="2025-12-04 17:58:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:59:02.196206845 +0000 UTC m=+1204.151567911" watchObservedRunningTime="2025-12-04 17:59:02.208368748 +0000 UTC m=+1204.163729784" Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.472878 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9t2z2"] Dec 04 17:59:02 crc kubenswrapper[4733]: I1204 17:59:02.884957 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:02 crc kubenswrapper[4733]: E1204 17:59:02.885137 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 17:59:02 crc kubenswrapper[4733]: E1204 17:59:02.885157 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 17:59:02 crc kubenswrapper[4733]: E1204 17:59:02.885215 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift podName:75da96fa-8079-442d-8993-de5658a23f91 nodeName:}" failed. No retries permitted until 2025-12-04 17:59:04.885196855 +0000 UTC m=+1206.840557891 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift") pod "swift-storage-0" (UID: "75da96fa-8079-442d-8993-de5658a23f91") : configmap "swift-ring-files" not found Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.183575 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9t2z2" event={"ID":"3590b958-1ae5-46f6-9664-d8efdec482b2","Type":"ContainerStarted","Data":"85dddfe4527b56482b7f41d7acf3b002ae0a114b74b004e23d9ad22bee68e155"} Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.240807 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-p2kxj"] Dec 04 17:59:03 crc kubenswrapper[4733]: E1204 17:59:03.241183 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b35785d-5a00-4554-b0ac-5585ca1c4816" containerName="mariadb-database-create" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.241203 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b35785d-5a00-4554-b0ac-5585ca1c4816" containerName="mariadb-database-create" Dec 04 17:59:03 crc kubenswrapper[4733]: E1204 17:59:03.241214 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b8dc93-1d10-4355-9d7f-ce423a3c11bc" containerName="mariadb-account-create-update" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.241221 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b8dc93-1d10-4355-9d7f-ce423a3c11bc" containerName="mariadb-account-create-update" Dec 04 17:59:03 crc kubenswrapper[4733]: E1204 17:59:03.241235 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a32f53d-9569-405a-9060-80e298dfc6ce" containerName="mariadb-database-create" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.241242 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a32f53d-9569-405a-9060-80e298dfc6ce" containerName="mariadb-database-create" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.241413 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b35785d-5a00-4554-b0ac-5585ca1c4816" containerName="mariadb-database-create" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.241433 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a32f53d-9569-405a-9060-80e298dfc6ce" containerName="mariadb-database-create" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.241447 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b8dc93-1d10-4355-9d7f-ce423a3c11bc" containerName="mariadb-account-create-update" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.241973 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.243785 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g46cp" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.255208 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.260457 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-p2kxj"] Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.298130 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-config-data\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.298220 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-db-sync-config-data\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.298348 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5nfd\" (UniqueName: \"kubernetes.io/projected/857cacd9-91d0-4a3a-92b8-73669fd75dfc-kube-api-access-n5nfd\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.298375 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-combined-ca-bundle\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.399360 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-db-sync-config-data\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.399492 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5nfd\" (UniqueName: \"kubernetes.io/projected/857cacd9-91d0-4a3a-92b8-73669fd75dfc-kube-api-access-n5nfd\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.399522 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-combined-ca-bundle\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.399545 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-config-data\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.404613 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-config-data\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.405064 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-combined-ca-bundle\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.405143 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-db-sync-config-data\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.419351 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5nfd\" (UniqueName: \"kubernetes.io/projected/857cacd9-91d0-4a3a-92b8-73669fd75dfc-kube-api-access-n5nfd\") pod \"glance-db-sync-p2kxj\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:03 crc kubenswrapper[4733]: I1204 17:59:03.558857 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:04 crc kubenswrapper[4733]: I1204 17:59:04.117430 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-p2kxj"] Dec 04 17:59:04 crc kubenswrapper[4733]: W1204 17:59:04.135541 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod857cacd9_91d0_4a3a_92b8_73669fd75dfc.slice/crio-e389e6b35e9b6f5026b9684947719ccb287706e0c6166b6f991e182f5960a3e8 WatchSource:0}: Error finding container e389e6b35e9b6f5026b9684947719ccb287706e0c6166b6f991e182f5960a3e8: Status 404 returned error can't find the container with id e389e6b35e9b6f5026b9684947719ccb287706e0c6166b6f991e182f5960a3e8 Dec 04 17:59:04 crc kubenswrapper[4733]: I1204 17:59:04.190397 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-p2kxj" event={"ID":"857cacd9-91d0-4a3a-92b8-73669fd75dfc","Type":"ContainerStarted","Data":"e389e6b35e9b6f5026b9684947719ccb287706e0c6166b6f991e182f5960a3e8"} Dec 04 17:59:04 crc kubenswrapper[4733]: I1204 17:59:04.412961 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 04 17:59:04 crc kubenswrapper[4733]: I1204 17:59:04.925613 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:04 crc kubenswrapper[4733]: E1204 17:59:04.925922 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 17:59:04 crc kubenswrapper[4733]: E1204 17:59:04.926111 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 17:59:04 crc kubenswrapper[4733]: E1204 17:59:04.926201 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift podName:75da96fa-8079-442d-8993-de5658a23f91 nodeName:}" failed. No retries permitted until 2025-12-04 17:59:08.926174233 +0000 UTC m=+1210.881535309 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift") pod "swift-storage-0" (UID: "75da96fa-8079-442d-8993-de5658a23f91") : configmap "swift-ring-files" not found Dec 04 17:59:08 crc kubenswrapper[4733]: I1204 17:59:08.239019 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9t2z2" event={"ID":"3590b958-1ae5-46f6-9664-d8efdec482b2","Type":"ContainerStarted","Data":"50af9429ee9c06ced96baa47e5ad56e3b3acb94ca909070661aefdb62d011c86"} Dec 04 17:59:08 crc kubenswrapper[4733]: I1204 17:59:08.354735 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9t2z2" podStartSLOduration=2.316002349 podStartE2EDuration="7.354715966s" podCreationTimestamp="2025-12-04 17:59:01 +0000 UTC" firstStartedPulling="2025-12-04 17:59:02.476489428 +0000 UTC m=+1204.431850474" lastFinishedPulling="2025-12-04 17:59:07.515203045 +0000 UTC m=+1209.470564091" observedRunningTime="2025-12-04 17:59:08.258063407 +0000 UTC m=+1210.213424473" watchObservedRunningTime="2025-12-04 17:59:08.354715966 +0000 UTC m=+1210.310077012" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.017404 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:09 crc kubenswrapper[4733]: E1204 17:59:09.017641 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 17:59:09 crc kubenswrapper[4733]: E1204 17:59:09.017671 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 17:59:09 crc kubenswrapper[4733]: E1204 17:59:09.017741 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift podName:75da96fa-8079-442d-8993-de5658a23f91 nodeName:}" failed. No retries permitted until 2025-12-04 17:59:17.017718735 +0000 UTC m=+1218.973079781 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift") pod "swift-storage-0" (UID: "75da96fa-8079-442d-8993-de5658a23f91") : configmap "swift-ring-files" not found Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.211131 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-j58xt" podUID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" containerName="ovn-controller" probeResult="failure" output=< Dec 04 17:59:09 crc kubenswrapper[4733]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 17:59:09 crc kubenswrapper[4733]: > Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.244413 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.249958 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.490525 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-j58xt-config-bmlvg"] Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.491560 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.495068 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.505896 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-j58xt-config-bmlvg"] Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.525578 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-scripts\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.525624 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-additional-scripts\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.525660 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-log-ovn\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.525684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run-ovn\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.525746 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49fls\" (UniqueName: \"kubernetes.io/projected/ae3198b4-c285-40d0-8e6f-10cc01c700f2-kube-api-access-49fls\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.525767 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.626915 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49fls\" (UniqueName: \"kubernetes.io/projected/ae3198b4-c285-40d0-8e6f-10cc01c700f2-kube-api-access-49fls\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.626963 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.627025 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-scripts\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.627044 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-additional-scripts\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.627079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-log-ovn\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.627110 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run-ovn\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.627394 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run-ovn\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.627422 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.627448 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-log-ovn\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.627908 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-additional-scripts\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.629021 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-scripts\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.644585 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49fls\" (UniqueName: \"kubernetes.io/projected/ae3198b4-c285-40d0-8e6f-10cc01c700f2-kube-api-access-49fls\") pod \"ovn-controller-j58xt-config-bmlvg\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:09 crc kubenswrapper[4733]: I1204 17:59:09.831824 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:10 crc kubenswrapper[4733]: I1204 17:59:10.189996 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:10 crc kubenswrapper[4733]: I1204 17:59:10.240284 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-rg88m"] Dec 04 17:59:10 crc kubenswrapper[4733]: I1204 17:59:10.240504 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="dnsmasq-dns" containerID="cri-o://46128b493a7ea8a8202efe06da8905c951a3b62403f1160fd1690a7f41d05ba3" gracePeriod=10 Dec 04 17:59:10 crc kubenswrapper[4733]: I1204 17:59:10.315323 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-j58xt-config-bmlvg"] Dec 04 17:59:10 crc kubenswrapper[4733]: W1204 17:59:10.320830 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae3198b4_c285_40d0_8e6f_10cc01c700f2.slice/crio-065e2cdb74ea69f5afd157bb42833814424767ad41e7d2bae61f09550afaf5f7 WatchSource:0}: Error finding container 065e2cdb74ea69f5afd157bb42833814424767ad41e7d2bae61f09550afaf5f7: Status 404 returned error can't find the container with id 065e2cdb74ea69f5afd157bb42833814424767ad41e7d2bae61f09550afaf5f7 Dec 04 17:59:11 crc kubenswrapper[4733]: I1204 17:59:11.289426 4733 generic.go:334] "Generic (PLEG): container finished" podID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerID="46128b493a7ea8a8202efe06da8905c951a3b62403f1160fd1690a7f41d05ba3" exitCode=0 Dec 04 17:59:11 crc kubenswrapper[4733]: I1204 17:59:11.289474 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" event={"ID":"a35bc564-4823-4bb3-9a04-7e1bdcc56a17","Type":"ContainerDied","Data":"46128b493a7ea8a8202efe06da8905c951a3b62403f1160fd1690a7f41d05ba3"} Dec 04 17:59:11 crc kubenswrapper[4733]: I1204 17:59:11.291962 4733 generic.go:334] "Generic (PLEG): container finished" podID="ae3198b4-c285-40d0-8e6f-10cc01c700f2" containerID="1cf76a2cd64147788487fcf9fe8b7751a8a4db7c081716cc42c264a1a160bb4f" exitCode=0 Dec 04 17:59:11 crc kubenswrapper[4733]: I1204 17:59:11.291997 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt-config-bmlvg" event={"ID":"ae3198b4-c285-40d0-8e6f-10cc01c700f2","Type":"ContainerDied","Data":"1cf76a2cd64147788487fcf9fe8b7751a8a4db7c081716cc42c264a1a160bb4f"} Dec 04 17:59:11 crc kubenswrapper[4733]: I1204 17:59:11.292019 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt-config-bmlvg" event={"ID":"ae3198b4-c285-40d0-8e6f-10cc01c700f2","Type":"ContainerStarted","Data":"065e2cdb74ea69f5afd157bb42833814424767ad41e7d2bae61f09550afaf5f7"} Dec 04 17:59:14 crc kubenswrapper[4733]: I1204 17:59:14.166428 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-j58xt" Dec 04 17:59:15 crc kubenswrapper[4733]: I1204 17:59:15.324712 4733 generic.go:334] "Generic (PLEG): container finished" podID="3590b958-1ae5-46f6-9664-d8efdec482b2" containerID="50af9429ee9c06ced96baa47e5ad56e3b3acb94ca909070661aefdb62d011c86" exitCode=0 Dec 04 17:59:15 crc kubenswrapper[4733]: I1204 17:59:15.324832 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9t2z2" event={"ID":"3590b958-1ae5-46f6-9664-d8efdec482b2","Type":"ContainerDied","Data":"50af9429ee9c06ced96baa47e5ad56e3b3acb94ca909070661aefdb62d011c86"} Dec 04 17:59:15 crc kubenswrapper[4733]: I1204 17:59:15.362069 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 17:59:15 crc kubenswrapper[4733]: I1204 17:59:15.362134 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 17:59:15 crc kubenswrapper[4733]: I1204 17:59:15.362184 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 17:59:15 crc kubenswrapper[4733]: I1204 17:59:15.362893 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f00595a749ca89ac07966111bfcf1cdeb8fcd3b039701228e808a5c55fac2e38"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 17:59:15 crc kubenswrapper[4733]: I1204 17:59:15.362966 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://f00595a749ca89ac07966111bfcf1cdeb8fcd3b039701228e808a5c55fac2e38" gracePeriod=600 Dec 04 17:59:16 crc kubenswrapper[4733]: I1204 17:59:16.339201 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="f00595a749ca89ac07966111bfcf1cdeb8fcd3b039701228e808a5c55fac2e38" exitCode=0 Dec 04 17:59:16 crc kubenswrapper[4733]: I1204 17:59:16.345751 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"f00595a749ca89ac07966111bfcf1cdeb8fcd3b039701228e808a5c55fac2e38"} Dec 04 17:59:16 crc kubenswrapper[4733]: I1204 17:59:16.345824 4733 scope.go:117] "RemoveContainer" containerID="245d5f7c913eadd0235bcdf1e7f003bb02e52afa3a7ef2fd2c9a9a3603961fd8" Dec 04 17:59:17 crc kubenswrapper[4733]: I1204 17:59:17.052836 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:17 crc kubenswrapper[4733]: I1204 17:59:17.061980 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"swift-storage-0\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " pod="openstack/swift-storage-0" Dec 04 17:59:17 crc kubenswrapper[4733]: I1204 17:59:17.325348 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 17:59:18 crc kubenswrapper[4733]: I1204 17:59:18.935101 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Dec 04 17:59:20 crc kubenswrapper[4733]: I1204 17:59:20.372131 4733 generic.go:334] "Generic (PLEG): container finished" podID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerID="9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab" exitCode=0 Dec 04 17:59:20 crc kubenswrapper[4733]: I1204 17:59:20.372264 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e605c62-64b4-4417-80bb-bc3387881f7a","Type":"ContainerDied","Data":"9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab"} Dec 04 17:59:23 crc kubenswrapper[4733]: I1204 17:59:23.935747 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Dec 04 17:59:24 crc kubenswrapper[4733]: E1204 17:59:24.004976 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63" Dec 04 17:59:24 crc kubenswrapper[4733]: E1204 17:59:24.005491 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n5nfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-p2kxj_openstack(857cacd9-91d0-4a3a-92b8-73669fd75dfc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 17:59:24 crc kubenswrapper[4733]: E1204 17:59:24.006987 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-p2kxj" podUID="857cacd9-91d0-4a3a-92b8-73669fd75dfc" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.070990 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.095857 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.120687 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158657 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49fls\" (UniqueName: \"kubernetes.io/projected/ae3198b4-c285-40d0-8e6f-10cc01c700f2-kube-api-access-49fls\") pod \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run-ovn\") pod \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158716 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-nb\") pod \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158752 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3590b958-1ae5-46f6-9664-d8efdec482b2-etc-swift\") pod \"3590b958-1ae5-46f6-9664-d8efdec482b2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158769 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7qs5\" (UniqueName: \"kubernetes.io/projected/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-kube-api-access-b7qs5\") pod \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158788 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-ring-data-devices\") pod \"3590b958-1ae5-46f6-9664-d8efdec482b2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158820 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-scripts\") pod \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158843 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-log-ovn\") pod \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158866 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-sb\") pod \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158884 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-config\") pod \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158903 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-additional-scripts\") pod \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158924 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-dns-svc\") pod \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\" (UID: \"a35bc564-4823-4bb3-9a04-7e1bdcc56a17\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158941 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-dispersionconf\") pod \"3590b958-1ae5-46f6-9664-d8efdec482b2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.158979 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-combined-ca-bundle\") pod \"3590b958-1ae5-46f6-9664-d8efdec482b2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.159018 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run\") pod \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\" (UID: \"ae3198b4-c285-40d0-8e6f-10cc01c700f2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.159040 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-scripts\") pod \"3590b958-1ae5-46f6-9664-d8efdec482b2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.159066 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqg8x\" (UniqueName: \"kubernetes.io/projected/3590b958-1ae5-46f6-9664-d8efdec482b2-kube-api-access-bqg8x\") pod \"3590b958-1ae5-46f6-9664-d8efdec482b2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.159092 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-swiftconf\") pod \"3590b958-1ae5-46f6-9664-d8efdec482b2\" (UID: \"3590b958-1ae5-46f6-9664-d8efdec482b2\") " Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.160307 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ae3198b4-c285-40d0-8e6f-10cc01c700f2" (UID: "ae3198b4-c285-40d0-8e6f-10cc01c700f2"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.160896 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "ae3198b4-c285-40d0-8e6f-10cc01c700f2" (UID: "ae3198b4-c285-40d0-8e6f-10cc01c700f2"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.161010 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3590b958-1ae5-46f6-9664-d8efdec482b2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3590b958-1ae5-46f6-9664-d8efdec482b2" (UID: "3590b958-1ae5-46f6-9664-d8efdec482b2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.162714 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3590b958-1ae5-46f6-9664-d8efdec482b2" (UID: "3590b958-1ae5-46f6-9664-d8efdec482b2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.162767 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ae3198b4-c285-40d0-8e6f-10cc01c700f2" (UID: "ae3198b4-c285-40d0-8e6f-10cc01c700f2"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.162879 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run" (OuterVolumeSpecName: "var-run") pod "ae3198b4-c285-40d0-8e6f-10cc01c700f2" (UID: "ae3198b4-c285-40d0-8e6f-10cc01c700f2"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.163177 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-scripts" (OuterVolumeSpecName: "scripts") pod "ae3198b4-c285-40d0-8e6f-10cc01c700f2" (UID: "ae3198b4-c285-40d0-8e6f-10cc01c700f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.168278 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae3198b4-c285-40d0-8e6f-10cc01c700f2-kube-api-access-49fls" (OuterVolumeSpecName: "kube-api-access-49fls") pod "ae3198b4-c285-40d0-8e6f-10cc01c700f2" (UID: "ae3198b4-c285-40d0-8e6f-10cc01c700f2"). InnerVolumeSpecName "kube-api-access-49fls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.183834 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3590b958-1ae5-46f6-9664-d8efdec482b2" (UID: "3590b958-1ae5-46f6-9664-d8efdec482b2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.183941 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3590b958-1ae5-46f6-9664-d8efdec482b2-kube-api-access-bqg8x" (OuterVolumeSpecName: "kube-api-access-bqg8x") pod "3590b958-1ae5-46f6-9664-d8efdec482b2" (UID: "3590b958-1ae5-46f6-9664-d8efdec482b2"). InnerVolumeSpecName "kube-api-access-bqg8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.184021 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-kube-api-access-b7qs5" (OuterVolumeSpecName: "kube-api-access-b7qs5") pod "a35bc564-4823-4bb3-9a04-7e1bdcc56a17" (UID: "a35bc564-4823-4bb3-9a04-7e1bdcc56a17"). InnerVolumeSpecName "kube-api-access-b7qs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.201475 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-scripts" (OuterVolumeSpecName: "scripts") pod "3590b958-1ae5-46f6-9664-d8efdec482b2" (UID: "3590b958-1ae5-46f6-9664-d8efdec482b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.227372 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3590b958-1ae5-46f6-9664-d8efdec482b2" (UID: "3590b958-1ae5-46f6-9664-d8efdec482b2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.228309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3590b958-1ae5-46f6-9664-d8efdec482b2" (UID: "3590b958-1ae5-46f6-9664-d8efdec482b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.230644 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a35bc564-4823-4bb3-9a04-7e1bdcc56a17" (UID: "a35bc564-4823-4bb3-9a04-7e1bdcc56a17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.240437 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a35bc564-4823-4bb3-9a04-7e1bdcc56a17" (UID: "a35bc564-4823-4bb3-9a04-7e1bdcc56a17"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.241546 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a35bc564-4823-4bb3-9a04-7e1bdcc56a17" (UID: "a35bc564-4823-4bb3-9a04-7e1bdcc56a17"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.251224 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-config" (OuterVolumeSpecName: "config") pod "a35bc564-4823-4bb3-9a04-7e1bdcc56a17" (UID: "a35bc564-4823-4bb3-9a04-7e1bdcc56a17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260471 4733 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260513 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260525 4733 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260536 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260546 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260556 4733 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae3198b4-c285-40d0-8e6f-10cc01c700f2-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260566 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260576 4733 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260586 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260599 4733 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260609 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3590b958-1ae5-46f6-9664-d8efdec482b2-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260619 4733 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3590b958-1ae5-46f6-9664-d8efdec482b2-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260631 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqg8x\" (UniqueName: \"kubernetes.io/projected/3590b958-1ae5-46f6-9664-d8efdec482b2-kube-api-access-bqg8x\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260644 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49fls\" (UniqueName: \"kubernetes.io/projected/ae3198b4-c285-40d0-8e6f-10cc01c700f2-kube-api-access-49fls\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260654 4733 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae3198b4-c285-40d0-8e6f-10cc01c700f2-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260663 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260672 4733 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3590b958-1ae5-46f6-9664-d8efdec482b2-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.260682 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7qs5\" (UniqueName: \"kubernetes.io/projected/a35bc564-4823-4bb3-9a04-7e1bdcc56a17-kube-api-access-b7qs5\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.407727 4733 generic.go:334] "Generic (PLEG): container finished" podID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerID="d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce" exitCode=0 Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.407783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f55e362-d59c-4269-92c3-d5ca014a2ef1","Type":"ContainerDied","Data":"d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce"} Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.425769 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.438093 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e605c62-64b4-4417-80bb-bc3387881f7a","Type":"ContainerStarted","Data":"d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d"} Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.438620 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.449147 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" event={"ID":"a35bc564-4823-4bb3-9a04-7e1bdcc56a17","Type":"ContainerDied","Data":"d2ca99acbf3e131ee876c40e3c794c60654fc8ca44f088b7896f57e01adb6e1e"} Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.449222 4733 scope.go:117] "RemoveContainer" containerID="46128b493a7ea8a8202efe06da8905c951a3b62403f1160fd1690a7f41d05ba3" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.449182 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.453314 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9t2z2" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.453343 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9t2z2" event={"ID":"3590b958-1ae5-46f6-9664-d8efdec482b2","Type":"ContainerDied","Data":"85dddfe4527b56482b7f41d7acf3b002ae0a114b74b004e23d9ad22bee68e155"} Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.453380 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85dddfe4527b56482b7f41d7acf3b002ae0a114b74b004e23d9ad22bee68e155" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.460987 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"f1e31ad24991110558bf25e49f4521223e1cb18d3a3ee43198b8eff935ecc2a3"} Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.465214 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt-config-bmlvg" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.465598 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt-config-bmlvg" event={"ID":"ae3198b4-c285-40d0-8e6f-10cc01c700f2","Type":"ContainerDied","Data":"065e2cdb74ea69f5afd157bb42833814424767ad41e7d2bae61f09550afaf5f7"} Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.465624 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="065e2cdb74ea69f5afd157bb42833814424767ad41e7d2bae61f09550afaf5f7" Dec 04 17:59:24 crc kubenswrapper[4733]: E1204 17:59:24.465810 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63\\\"\"" pod="openstack/glance-db-sync-p2kxj" podUID="857cacd9-91d0-4a3a-92b8-73669fd75dfc" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.482525 4733 scope.go:117] "RemoveContainer" containerID="b7d68bd6f9f6d1d744eb35792427c56d354195501d227587210172786e20992a" Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.524906 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-rg88m"] Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.537871 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-rg88m"] Dec 04 17:59:24 crc kubenswrapper[4733]: I1204 17:59:24.568365 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.811671367 podStartE2EDuration="1m32.568348708s" podCreationTimestamp="2025-12-04 17:57:52 +0000 UTC" firstStartedPulling="2025-12-04 17:57:55.08966291 +0000 UTC m=+1137.045023956" lastFinishedPulling="2025-12-04 17:58:44.846340251 +0000 UTC m=+1186.801701297" observedRunningTime="2025-12-04 17:59:24.561973748 +0000 UTC m=+1226.517334794" watchObservedRunningTime="2025-12-04 17:59:24.568348708 +0000 UTC m=+1226.523709754" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.243939 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-j58xt-config-bmlvg"] Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.250914 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-j58xt-config-bmlvg"] Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.335319 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-j58xt-config-8qb2d"] Dec 04 17:59:25 crc kubenswrapper[4733]: E1204 17:59:25.335685 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="init" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.335721 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="init" Dec 04 17:59:25 crc kubenswrapper[4733]: E1204 17:59:25.335742 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3198b4-c285-40d0-8e6f-10cc01c700f2" containerName="ovn-config" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.335753 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3198b4-c285-40d0-8e6f-10cc01c700f2" containerName="ovn-config" Dec 04 17:59:25 crc kubenswrapper[4733]: E1204 17:59:25.335780 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="dnsmasq-dns" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.335850 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="dnsmasq-dns" Dec 04 17:59:25 crc kubenswrapper[4733]: E1204 17:59:25.335866 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3590b958-1ae5-46f6-9664-d8efdec482b2" containerName="swift-ring-rebalance" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.335876 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3590b958-1ae5-46f6-9664-d8efdec482b2" containerName="swift-ring-rebalance" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.336059 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="dnsmasq-dns" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.336077 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3590b958-1ae5-46f6-9664-d8efdec482b2" containerName="swift-ring-rebalance" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.336091 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3198b4-c285-40d0-8e6f-10cc01c700f2" containerName="ovn-config" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.336640 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.338722 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.346489 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-j58xt-config-8qb2d"] Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.474548 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f55e362-d59c-4269-92c3-d5ca014a2ef1","Type":"ContainerStarted","Data":"98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25"} Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.475660 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.477167 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"fd17271bff52c7522d9ce0902111b6a3f9ff36c957ced644ec80c5590f8cacb9"} Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.485718 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run-ovn\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.485776 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.485901 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-additional-scripts\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.487031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-scripts\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.487204 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5crpk\" (UniqueName: \"kubernetes.io/projected/f356cda2-2da9-4380-bdcf-368aebdbd202-kube-api-access-5crpk\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.487275 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-log-ovn\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.516260 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371944.33854 podStartE2EDuration="1m32.516237362s" podCreationTimestamp="2025-12-04 17:57:53 +0000 UTC" firstStartedPulling="2025-12-04 17:57:55.191551209 +0000 UTC m=+1137.146912255" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:59:25.507475699 +0000 UTC m=+1227.462836765" watchObservedRunningTime="2025-12-04 17:59:25.516237362 +0000 UTC m=+1227.471598408" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.589237 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-scripts\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.589378 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5crpk\" (UniqueName: \"kubernetes.io/projected/f356cda2-2da9-4380-bdcf-368aebdbd202-kube-api-access-5crpk\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.589459 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-log-ovn\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.589562 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run-ovn\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.589718 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.589862 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-log-ovn\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.589877 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-additional-scripts\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.590156 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run-ovn\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.590227 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.590670 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-additional-scripts\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.591653 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-scripts\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.606862 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5crpk\" (UniqueName: \"kubernetes.io/projected/f356cda2-2da9-4380-bdcf-368aebdbd202-kube-api-access-5crpk\") pod \"ovn-controller-j58xt-config-8qb2d\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:25 crc kubenswrapper[4733]: I1204 17:59:25.652090 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:26 crc kubenswrapper[4733]: I1204 17:59:26.156177 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-j58xt-config-8qb2d"] Dec 04 17:59:26 crc kubenswrapper[4733]: W1204 17:59:26.173930 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf356cda2_2da9_4380_bdcf_368aebdbd202.slice/crio-2d6c27d82390f2e52d97a41da063b7218f5b998a60dc0372a1e2d469c6f4bb1e WatchSource:0}: Error finding container 2d6c27d82390f2e52d97a41da063b7218f5b998a60dc0372a1e2d469c6f4bb1e: Status 404 returned error can't find the container with id 2d6c27d82390f2e52d97a41da063b7218f5b998a60dc0372a1e2d469c6f4bb1e Dec 04 17:59:26 crc kubenswrapper[4733]: I1204 17:59:26.347900 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" path="/var/lib/kubelet/pods/a35bc564-4823-4bb3-9a04-7e1bdcc56a17/volumes" Dec 04 17:59:26 crc kubenswrapper[4733]: I1204 17:59:26.348961 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae3198b4-c285-40d0-8e6f-10cc01c700f2" path="/var/lib/kubelet/pods/ae3198b4-c285-40d0-8e6f-10cc01c700f2/volumes" Dec 04 17:59:26 crc kubenswrapper[4733]: I1204 17:59:26.495663 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt-config-8qb2d" event={"ID":"f356cda2-2da9-4380-bdcf-368aebdbd202","Type":"ContainerStarted","Data":"2d6c27d82390f2e52d97a41da063b7218f5b998a60dc0372a1e2d469c6f4bb1e"} Dec 04 17:59:27 crc kubenswrapper[4733]: I1204 17:59:27.533502 4733 generic.go:334] "Generic (PLEG): container finished" podID="f356cda2-2da9-4380-bdcf-368aebdbd202" containerID="cf6ccb65296ed2b05152275d1569159933dbce9caea4d4b0634f3470dd6b2429" exitCode=0 Dec 04 17:59:27 crc kubenswrapper[4733]: I1204 17:59:27.533596 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt-config-8qb2d" event={"ID":"f356cda2-2da9-4380-bdcf-368aebdbd202","Type":"ContainerDied","Data":"cf6ccb65296ed2b05152275d1569159933dbce9caea4d4b0634f3470dd6b2429"} Dec 04 17:59:27 crc kubenswrapper[4733]: I1204 17:59:27.549118 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d"} Dec 04 17:59:27 crc kubenswrapper[4733]: I1204 17:59:27.549178 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04"} Dec 04 17:59:27 crc kubenswrapper[4733]: I1204 17:59:27.549204 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0"} Dec 04 17:59:28 crc kubenswrapper[4733]: I1204 17:59:28.564086 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018"} Dec 04 17:59:28 crc kubenswrapper[4733]: I1204 17:59:28.936273 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-rg88m" podUID="a35bc564-4823-4bb3-9a04-7e1bdcc56a17" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.085452 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243259 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-additional-scripts\") pod \"f356cda2-2da9-4380-bdcf-368aebdbd202\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243298 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-scripts\") pod \"f356cda2-2da9-4380-bdcf-368aebdbd202\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243330 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run-ovn\") pod \"f356cda2-2da9-4380-bdcf-368aebdbd202\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243381 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5crpk\" (UniqueName: \"kubernetes.io/projected/f356cda2-2da9-4380-bdcf-368aebdbd202-kube-api-access-5crpk\") pod \"f356cda2-2da9-4380-bdcf-368aebdbd202\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243447 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run\") pod \"f356cda2-2da9-4380-bdcf-368aebdbd202\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243529 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-log-ovn\") pod \"f356cda2-2da9-4380-bdcf-368aebdbd202\" (UID: \"f356cda2-2da9-4380-bdcf-368aebdbd202\") " Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243733 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run" (OuterVolumeSpecName: "var-run") pod "f356cda2-2da9-4380-bdcf-368aebdbd202" (UID: "f356cda2-2da9-4380-bdcf-368aebdbd202"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243808 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f356cda2-2da9-4380-bdcf-368aebdbd202" (UID: "f356cda2-2da9-4380-bdcf-368aebdbd202"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.243806 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f356cda2-2da9-4380-bdcf-368aebdbd202" (UID: "f356cda2-2da9-4380-bdcf-368aebdbd202"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.244162 4733 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.244176 4733 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.244187 4733 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f356cda2-2da9-4380-bdcf-368aebdbd202-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.244490 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f356cda2-2da9-4380-bdcf-368aebdbd202" (UID: "f356cda2-2da9-4380-bdcf-368aebdbd202"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.244504 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-scripts" (OuterVolumeSpecName: "scripts") pod "f356cda2-2da9-4380-bdcf-368aebdbd202" (UID: "f356cda2-2da9-4380-bdcf-368aebdbd202"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.250175 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f356cda2-2da9-4380-bdcf-368aebdbd202-kube-api-access-5crpk" (OuterVolumeSpecName: "kube-api-access-5crpk") pod "f356cda2-2da9-4380-bdcf-368aebdbd202" (UID: "f356cda2-2da9-4380-bdcf-368aebdbd202"). InnerVolumeSpecName "kube-api-access-5crpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.345699 4733 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.346004 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356cda2-2da9-4380-bdcf-368aebdbd202-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.346014 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5crpk\" (UniqueName: \"kubernetes.io/projected/f356cda2-2da9-4380-bdcf-368aebdbd202-kube-api-access-5crpk\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.573783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt-config-8qb2d" event={"ID":"f356cda2-2da9-4380-bdcf-368aebdbd202","Type":"ContainerDied","Data":"2d6c27d82390f2e52d97a41da063b7218f5b998a60dc0372a1e2d469c6f4bb1e"} Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.573827 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt-config-8qb2d" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.573840 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d6c27d82390f2e52d97a41da063b7218f5b998a60dc0372a1e2d469c6f4bb1e" Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.584025 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95"} Dec 04 17:59:29 crc kubenswrapper[4733]: I1204 17:59:29.584061 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc"} Dec 04 17:59:30 crc kubenswrapper[4733]: I1204 17:59:30.166019 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-j58xt-config-8qb2d"] Dec 04 17:59:30 crc kubenswrapper[4733]: I1204 17:59:30.173470 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-j58xt-config-8qb2d"] Dec 04 17:59:30 crc kubenswrapper[4733]: I1204 17:59:30.344244 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f356cda2-2da9-4380-bdcf-368aebdbd202" path="/var/lib/kubelet/pods/f356cda2-2da9-4380-bdcf-368aebdbd202/volumes" Dec 04 17:59:30 crc kubenswrapper[4733]: I1204 17:59:30.595597 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca"} Dec 04 17:59:30 crc kubenswrapper[4733]: I1204 17:59:30.597032 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87"} Dec 04 17:59:31 crc kubenswrapper[4733]: I1204 17:59:31.609605 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda"} Dec 04 17:59:31 crc kubenswrapper[4733]: I1204 17:59:31.609950 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd"} Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.626009 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c"} Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.626349 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c"} Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.626360 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0"} Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.626369 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828"} Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.626377 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerStarted","Data":"33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f"} Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.675505 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=26.035383908 podStartE2EDuration="32.675482847s" podCreationTimestamp="2025-12-04 17:59:00 +0000 UTC" firstStartedPulling="2025-12-04 17:59:24.439648585 +0000 UTC m=+1226.395009631" lastFinishedPulling="2025-12-04 17:59:31.079747524 +0000 UTC m=+1233.035108570" observedRunningTime="2025-12-04 17:59:32.670016103 +0000 UTC m=+1234.625377169" watchObservedRunningTime="2025-12-04 17:59:32.675482847 +0000 UTC m=+1234.630843893" Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.974999 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-ptkcg"] Dec 04 17:59:32 crc kubenswrapper[4733]: E1204 17:59:32.975740 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f356cda2-2da9-4380-bdcf-368aebdbd202" containerName="ovn-config" Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.975872 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f356cda2-2da9-4380-bdcf-368aebdbd202" containerName="ovn-config" Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.976220 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f356cda2-2da9-4380-bdcf-368aebdbd202" containerName="ovn-config" Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.977542 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.979824 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 04 17:59:32 crc kubenswrapper[4733]: I1204 17:59:32.999491 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-ptkcg"] Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.099534 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-svc\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.099591 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.099613 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4ws6\" (UniqueName: \"kubernetes.io/projected/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-kube-api-access-j4ws6\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.099773 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.099932 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.100005 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-config\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.201820 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-svc\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.201899 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.201927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4ws6\" (UniqueName: \"kubernetes.io/projected/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-kube-api-access-j4ws6\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.201988 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.202042 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.202069 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-config\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.203027 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-svc\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.203116 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-config\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.203125 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-swift-storage-0\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.203311 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-nb\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.204575 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-sb\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.263848 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4ws6\" (UniqueName: \"kubernetes.io/projected/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-kube-api-access-j4ws6\") pod \"dnsmasq-dns-779c5847bc-ptkcg\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.292775 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:33 crc kubenswrapper[4733]: I1204 17:59:33.747009 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-ptkcg"] Dec 04 17:59:33 crc kubenswrapper[4733]: W1204 17:59:33.751691 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fdc02f1_40d7_43a1_9869_7815e0e6d2db.slice/crio-a84adfb68199df745f8e243aeda49fb2c1210b88bba85d44083c058b1bb29308 WatchSource:0}: Error finding container a84adfb68199df745f8e243aeda49fb2c1210b88bba85d44083c058b1bb29308: Status 404 returned error can't find the container with id a84adfb68199df745f8e243aeda49fb2c1210b88bba85d44083c058b1bb29308 Dec 04 17:59:34 crc kubenswrapper[4733]: I1204 17:59:34.577019 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 17:59:34 crc kubenswrapper[4733]: I1204 17:59:34.642635 4733 generic.go:334] "Generic (PLEG): container finished" podID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerID="89c5383d6b95f43e1f6d91b8ef881a2cf07c6e3ee81c158f06ced81158b76bd4" exitCode=0 Dec 04 17:59:34 crc kubenswrapper[4733]: I1204 17:59:34.642688 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" event={"ID":"5fdc02f1-40d7-43a1-9869-7815e0e6d2db","Type":"ContainerDied","Data":"89c5383d6b95f43e1f6d91b8ef881a2cf07c6e3ee81c158f06ced81158b76bd4"} Dec 04 17:59:34 crc kubenswrapper[4733]: I1204 17:59:34.642717 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" event={"ID":"5fdc02f1-40d7-43a1-9869-7815e0e6d2db","Type":"ContainerStarted","Data":"a84adfb68199df745f8e243aeda49fb2c1210b88bba85d44083c058b1bb29308"} Dec 04 17:59:34 crc kubenswrapper[4733]: I1204 17:59:34.697700 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 17:59:35 crc kubenswrapper[4733]: I1204 17:59:35.655018 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" event={"ID":"5fdc02f1-40d7-43a1-9869-7815e0e6d2db","Type":"ContainerStarted","Data":"8d980a83191192ad31b4a99fa1e5f80399e038222d823f299b5b3eac1a730fda"} Dec 04 17:59:35 crc kubenswrapper[4733]: I1204 17:59:35.655254 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:35 crc kubenswrapper[4733]: I1204 17:59:35.676897 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" podStartSLOduration=3.676880334 podStartE2EDuration="3.676880334s" podCreationTimestamp="2025-12-04 17:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:59:35.673579297 +0000 UTC m=+1237.628940413" watchObservedRunningTime="2025-12-04 17:59:35.676880334 +0000 UTC m=+1237.632241380" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.690958 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-76hpp"] Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.692302 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.698291 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-2aa9-account-create-update-dd8zp"] Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.699478 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.701108 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.713251 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2aa9-account-create-update-dd8zp"] Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.724015 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-76hpp"] Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.775878 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41470d79-78b1-44bb-a10f-9f169d2d9e17-operator-scripts\") pod \"barbican-db-create-76hpp\" (UID: \"41470d79-78b1-44bb-a10f-9f169d2d9e17\") " pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.775990 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55hnn\" (UniqueName: \"kubernetes.io/projected/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-kube-api-access-55hnn\") pod \"barbican-2aa9-account-create-update-dd8zp\" (UID: \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\") " pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.776034 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-operator-scripts\") pod \"barbican-2aa9-account-create-update-dd8zp\" (UID: \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\") " pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.776060 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t66sb\" (UniqueName: \"kubernetes.io/projected/41470d79-78b1-44bb-a10f-9f169d2d9e17-kube-api-access-t66sb\") pod \"barbican-db-create-76hpp\" (UID: \"41470d79-78b1-44bb-a10f-9f169d2d9e17\") " pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.793362 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-q4prp"] Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.794731 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.801865 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-q4prp"] Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.877089 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwdgj\" (UniqueName: \"kubernetes.io/projected/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-kube-api-access-dwdgj\") pod \"cinder-db-create-q4prp\" (UID: \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\") " pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.877136 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41470d79-78b1-44bb-a10f-9f169d2d9e17-operator-scripts\") pod \"barbican-db-create-76hpp\" (UID: \"41470d79-78b1-44bb-a10f-9f169d2d9e17\") " pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.877367 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-operator-scripts\") pod \"cinder-db-create-q4prp\" (UID: \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\") " pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.877417 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55hnn\" (UniqueName: \"kubernetes.io/projected/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-kube-api-access-55hnn\") pod \"barbican-2aa9-account-create-update-dd8zp\" (UID: \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\") " pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.877455 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-operator-scripts\") pod \"barbican-2aa9-account-create-update-dd8zp\" (UID: \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\") " pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.877495 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t66sb\" (UniqueName: \"kubernetes.io/projected/41470d79-78b1-44bb-a10f-9f169d2d9e17-kube-api-access-t66sb\") pod \"barbican-db-create-76hpp\" (UID: \"41470d79-78b1-44bb-a10f-9f169d2d9e17\") " pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.878370 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41470d79-78b1-44bb-a10f-9f169d2d9e17-operator-scripts\") pod \"barbican-db-create-76hpp\" (UID: \"41470d79-78b1-44bb-a10f-9f169d2d9e17\") " pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.878973 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-operator-scripts\") pod \"barbican-2aa9-account-create-update-dd8zp\" (UID: \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\") " pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.907294 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55hnn\" (UniqueName: \"kubernetes.io/projected/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-kube-api-access-55hnn\") pod \"barbican-2aa9-account-create-update-dd8zp\" (UID: \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\") " pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.921332 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t66sb\" (UniqueName: \"kubernetes.io/projected/41470d79-78b1-44bb-a10f-9f169d2d9e17-kube-api-access-t66sb\") pod \"barbican-db-create-76hpp\" (UID: \"41470d79-78b1-44bb-a10f-9f169d2d9e17\") " pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.953252 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-7119-account-create-update-k2cvp"] Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.954247 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.957200 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.971482 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7119-account-create-update-k2cvp"] Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.978650 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwdgj\" (UniqueName: \"kubernetes.io/projected/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-kube-api-access-dwdgj\") pod \"cinder-db-create-q4prp\" (UID: \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\") " pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.978711 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-operator-scripts\") pod \"cinder-db-create-q4prp\" (UID: \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\") " pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:36 crc kubenswrapper[4733]: I1204 17:59:36.979542 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-operator-scripts\") pod \"cinder-db-create-q4prp\" (UID: \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\") " pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.009385 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwdgj\" (UniqueName: \"kubernetes.io/projected/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-kube-api-access-dwdgj\") pod \"cinder-db-create-q4prp\" (UID: \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\") " pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.021579 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.032479 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.044104 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-5vq76"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.047372 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.053695 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5vq76"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.089841 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsjnp\" (UniqueName: \"kubernetes.io/projected/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-kube-api-access-nsjnp\") pod \"cinder-7119-account-create-update-k2cvp\" (UID: \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\") " pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.089910 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-operator-scripts\") pod \"cinder-7119-account-create-update-k2cvp\" (UID: \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\") " pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.109328 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.140747 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-27f6z"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.141772 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.144572 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.154894 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.155139 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r9kr4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.176120 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.178522 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-27f6z"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.224490 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsjnp\" (UniqueName: \"kubernetes.io/projected/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-kube-api-access-nsjnp\") pod \"cinder-7119-account-create-update-k2cvp\" (UID: \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\") " pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.224835 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj87v\" (UniqueName: \"kubernetes.io/projected/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-kube-api-access-cj87v\") pod \"neutron-db-create-5vq76\" (UID: \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\") " pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.226721 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-operator-scripts\") pod \"cinder-7119-account-create-update-k2cvp\" (UID: \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\") " pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.227011 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-operator-scripts\") pod \"neutron-db-create-5vq76\" (UID: \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\") " pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.227927 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-operator-scripts\") pod \"cinder-7119-account-create-update-k2cvp\" (UID: \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\") " pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.241394 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-17fd-account-create-update-ct9z4"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.242528 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.252574 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.273863 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-17fd-account-create-update-ct9z4"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.286602 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsjnp\" (UniqueName: \"kubernetes.io/projected/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-kube-api-access-nsjnp\") pod \"cinder-7119-account-create-update-k2cvp\" (UID: \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\") " pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.293052 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.330039 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj87v\" (UniqueName: \"kubernetes.io/projected/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-kube-api-access-cj87v\") pod \"neutron-db-create-5vq76\" (UID: \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\") " pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.330142 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-combined-ca-bundle\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.330203 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-operator-scripts\") pod \"neutron-db-create-5vq76\" (UID: \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\") " pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.330224 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-config-data\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.330242 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jslsk\" (UniqueName: \"kubernetes.io/projected/37aa673f-32a7-43ce-b1b2-a7f02f46c485-kube-api-access-jslsk\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.331365 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-operator-scripts\") pod \"neutron-db-create-5vq76\" (UID: \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\") " pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.349273 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj87v\" (UniqueName: \"kubernetes.io/projected/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-kube-api-access-cj87v\") pod \"neutron-db-create-5vq76\" (UID: \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\") " pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.432984 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-config-data\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.433269 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jslsk\" (UniqueName: \"kubernetes.io/projected/37aa673f-32a7-43ce-b1b2-a7f02f46c485-kube-api-access-jslsk\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.433299 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30a71d3-ad60-465b-88f2-e642c701f598-operator-scripts\") pod \"neutron-17fd-account-create-update-ct9z4\" (UID: \"c30a71d3-ad60-465b-88f2-e642c701f598\") " pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.433338 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gszfv\" (UniqueName: \"kubernetes.io/projected/c30a71d3-ad60-465b-88f2-e642c701f598-kube-api-access-gszfv\") pod \"neutron-17fd-account-create-update-ct9z4\" (UID: \"c30a71d3-ad60-465b-88f2-e642c701f598\") " pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.433418 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-combined-ca-bundle\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.436388 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.439756 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-config-data\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.443300 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-combined-ca-bundle\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.466410 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jslsk\" (UniqueName: \"kubernetes.io/projected/37aa673f-32a7-43ce-b1b2-a7f02f46c485-kube-api-access-jslsk\") pod \"keystone-db-sync-27f6z\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.536409 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30a71d3-ad60-465b-88f2-e642c701f598-operator-scripts\") pod \"neutron-17fd-account-create-update-ct9z4\" (UID: \"c30a71d3-ad60-465b-88f2-e642c701f598\") " pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.536510 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gszfv\" (UniqueName: \"kubernetes.io/projected/c30a71d3-ad60-465b-88f2-e642c701f598-kube-api-access-gszfv\") pod \"neutron-17fd-account-create-update-ct9z4\" (UID: \"c30a71d3-ad60-465b-88f2-e642c701f598\") " pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.537100 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30a71d3-ad60-465b-88f2-e642c701f598-operator-scripts\") pod \"neutron-17fd-account-create-update-ct9z4\" (UID: \"c30a71d3-ad60-465b-88f2-e642c701f598\") " pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.551665 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gszfv\" (UniqueName: \"kubernetes.io/projected/c30a71d3-ad60-465b-88f2-e642c701f598-kube-api-access-gszfv\") pod \"neutron-17fd-account-create-update-ct9z4\" (UID: \"c30a71d3-ad60-465b-88f2-e642c701f598\") " pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.556460 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.577869 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-76hpp"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.590617 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.673566 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2aa9-account-create-update-dd8zp"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.680281 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-76hpp" event={"ID":"41470d79-78b1-44bb-a10f-9f169d2d9e17","Type":"ContainerStarted","Data":"bd56e6391be736cabdb0954cfab41f5d17c639bb1a0e07346bac10bbc89d15c4"} Dec 04 17:59:37 crc kubenswrapper[4733]: W1204 17:59:37.680378 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea7a064e_d1c8_4fd6_8755_f2e0b0e999e3.slice/crio-2b26d0391fe204d8a747bc3820108c9a015e516ccd97473b1229e6efe64d58b4 WatchSource:0}: Error finding container 2b26d0391fe204d8a747bc3820108c9a015e516ccd97473b1229e6efe64d58b4: Status 404 returned error can't find the container with id 2b26d0391fe204d8a747bc3820108c9a015e516ccd97473b1229e6efe64d58b4 Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.723936 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-q4prp"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.793961 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7119-account-create-update-k2cvp"] Dec 04 17:59:37 crc kubenswrapper[4733]: I1204 17:59:37.956929 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5vq76"] Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.070646 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-27f6z"] Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.148505 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-17fd-account-create-update-ct9z4"] Dec 04 17:59:38 crc kubenswrapper[4733]: W1204 17:59:38.157491 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc30a71d3_ad60_465b_88f2_e642c701f598.slice/crio-aeff95349210fc793d2e056be60182b99092a0ad9242487190cdd006ec769268 WatchSource:0}: Error finding container aeff95349210fc793d2e056be60182b99092a0ad9242487190cdd006ec769268: Status 404 returned error can't find the container with id aeff95349210fc793d2e056be60182b99092a0ad9242487190cdd006ec769268 Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.725610 4733 generic.go:334] "Generic (PLEG): container finished" podID="41470d79-78b1-44bb-a10f-9f169d2d9e17" containerID="9fc2b7a8967966a8f3ff5322d84258451c9bdd69b827ca292e5e59c5d79425d0" exitCode=0 Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.725692 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-76hpp" event={"ID":"41470d79-78b1-44bb-a10f-9f169d2d9e17","Type":"ContainerDied","Data":"9fc2b7a8967966a8f3ff5322d84258451c9bdd69b827ca292e5e59c5d79425d0"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.735006 4733 generic.go:334] "Generic (PLEG): container finished" podID="ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3" containerID="45806a9fe665f29c9726301d1f3e4806eb4602973959b29e42f98d3c93819ee4" exitCode=0 Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.735227 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2aa9-account-create-update-dd8zp" event={"ID":"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3","Type":"ContainerDied","Data":"45806a9fe665f29c9726301d1f3e4806eb4602973959b29e42f98d3c93819ee4"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.735257 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2aa9-account-create-update-dd8zp" event={"ID":"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3","Type":"ContainerStarted","Data":"2b26d0391fe204d8a747bc3820108c9a015e516ccd97473b1229e6efe64d58b4"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.736820 4733 generic.go:334] "Generic (PLEG): container finished" podID="d7fdf3ca-8982-4da7-8b15-1647ddb2fef3" containerID="ab9aabca56e540a2cedf02be5dda5c3c15082cf0c0d6c2af3a21c506250f6bcb" exitCode=0 Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.736885 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-q4prp" event={"ID":"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3","Type":"ContainerDied","Data":"ab9aabca56e540a2cedf02be5dda5c3c15082cf0c0d6c2af3a21c506250f6bcb"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.736915 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-q4prp" event={"ID":"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3","Type":"ContainerStarted","Data":"4ac312153d05e66c86ba777292d474b2e463be281bd0aac112fda771aedcb917"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.739729 4733 generic.go:334] "Generic (PLEG): container finished" podID="0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6" containerID="7c07b0d7d016fce0dca335c978df6940bc979f89cde5956c8726ce39d20debad" exitCode=0 Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.739877 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7119-account-create-update-k2cvp" event={"ID":"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6","Type":"ContainerDied","Data":"7c07b0d7d016fce0dca335c978df6940bc979f89cde5956c8726ce39d20debad"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.739905 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7119-account-create-update-k2cvp" event={"ID":"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6","Type":"ContainerStarted","Data":"1a6565899dd98f4da2fcbc0839fea8104b651464cc611a59289961e5f1946a43"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.745206 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-27f6z" event={"ID":"37aa673f-32a7-43ce-b1b2-a7f02f46c485","Type":"ContainerStarted","Data":"ba9e9e214d9e6c62c29b3b6c9ba0a43ad21db01046d3220beb84eb447fa7b420"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.747234 4733 generic.go:334] "Generic (PLEG): container finished" podID="40ea0fb9-d583-4ea8-8bb4-691e0f8606d3" containerID="70c93568467ddeaf24343edeb7cc5662dba7fac3fe6fb7f124ae21b60667977d" exitCode=0 Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.747311 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5vq76" event={"ID":"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3","Type":"ContainerDied","Data":"70c93568467ddeaf24343edeb7cc5662dba7fac3fe6fb7f124ae21b60667977d"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.747338 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5vq76" event={"ID":"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3","Type":"ContainerStarted","Data":"4ab907e76e224d156769209cdd2c3ffb5091c69dff5e6144c36da15b2900e76e"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.767850 4733 generic.go:334] "Generic (PLEG): container finished" podID="c30a71d3-ad60-465b-88f2-e642c701f598" containerID="beeeaa11c7814632a3f71a9a39055495a734228f830097d6a2f2043823fae14a" exitCode=0 Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.767911 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-17fd-account-create-update-ct9z4" event={"ID":"c30a71d3-ad60-465b-88f2-e642c701f598","Type":"ContainerDied","Data":"beeeaa11c7814632a3f71a9a39055495a734228f830097d6a2f2043823fae14a"} Dec 04 17:59:38 crc kubenswrapper[4733]: I1204 17:59:38.767949 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-17fd-account-create-update-ct9z4" event={"ID":"c30a71d3-ad60-465b-88f2-e642c701f598","Type":"ContainerStarted","Data":"aeff95349210fc793d2e056be60182b99092a0ad9242487190cdd006ec769268"} Dec 04 17:59:39 crc kubenswrapper[4733]: I1204 17:59:39.778270 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-p2kxj" event={"ID":"857cacd9-91d0-4a3a-92b8-73669fd75dfc","Type":"ContainerStarted","Data":"7daea43ad01541be1c71ecfc6cbf37ab9bcaf5bedff0f422646929d973829b99"} Dec 04 17:59:39 crc kubenswrapper[4733]: I1204 17:59:39.811210 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-p2kxj" podStartSLOduration=2.164483849 podStartE2EDuration="36.811185293s" podCreationTimestamp="2025-12-04 17:59:03 +0000 UTC" firstStartedPulling="2025-12-04 17:59:04.137996536 +0000 UTC m=+1206.093357592" lastFinishedPulling="2025-12-04 17:59:38.78469797 +0000 UTC m=+1240.740059036" observedRunningTime="2025-12-04 17:59:39.799301 +0000 UTC m=+1241.754662046" watchObservedRunningTime="2025-12-04 17:59:39.811185293 +0000 UTC m=+1241.766546349" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.195908 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.282080 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t66sb\" (UniqueName: \"kubernetes.io/projected/41470d79-78b1-44bb-a10f-9f169d2d9e17-kube-api-access-t66sb\") pod \"41470d79-78b1-44bb-a10f-9f169d2d9e17\" (UID: \"41470d79-78b1-44bb-a10f-9f169d2d9e17\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.282245 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41470d79-78b1-44bb-a10f-9f169d2d9e17-operator-scripts\") pod \"41470d79-78b1-44bb-a10f-9f169d2d9e17\" (UID: \"41470d79-78b1-44bb-a10f-9f169d2d9e17\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.282847 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41470d79-78b1-44bb-a10f-9f169d2d9e17-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "41470d79-78b1-44bb-a10f-9f169d2d9e17" (UID: "41470d79-78b1-44bb-a10f-9f169d2d9e17"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.288480 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41470d79-78b1-44bb-a10f-9f169d2d9e17-kube-api-access-t66sb" (OuterVolumeSpecName: "kube-api-access-t66sb") pod "41470d79-78b1-44bb-a10f-9f169d2d9e17" (UID: "41470d79-78b1-44bb-a10f-9f169d2d9e17"). InnerVolumeSpecName "kube-api-access-t66sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.345775 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.367308 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.398008 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t66sb\" (UniqueName: \"kubernetes.io/projected/41470d79-78b1-44bb-a10f-9f169d2d9e17-kube-api-access-t66sb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.398044 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41470d79-78b1-44bb-a10f-9f169d2d9e17-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.412987 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.428931 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.432142 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.498946 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-operator-scripts\") pod \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\" (UID: \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.499217 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-operator-scripts\") pod \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\" (UID: \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.499385 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsjnp\" (UniqueName: \"kubernetes.io/projected/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-kube-api-access-nsjnp\") pod \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\" (UID: \"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.499511 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj87v\" (UniqueName: \"kubernetes.io/projected/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-kube-api-access-cj87v\") pod \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\" (UID: \"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.499417 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6" (UID: "0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.499678 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "40ea0fb9-d583-4ea8-8bb4-691e0f8606d3" (UID: "40ea0fb9-d583-4ea8-8bb4-691e0f8606d3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.500163 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.500235 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.503333 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-kube-api-access-cj87v" (OuterVolumeSpecName: "kube-api-access-cj87v") pod "40ea0fb9-d583-4ea8-8bb4-691e0f8606d3" (UID: "40ea0fb9-d583-4ea8-8bb4-691e0f8606d3"). InnerVolumeSpecName "kube-api-access-cj87v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.511921 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-kube-api-access-nsjnp" (OuterVolumeSpecName: "kube-api-access-nsjnp") pod "0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6" (UID: "0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6"). InnerVolumeSpecName "kube-api-access-nsjnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.601512 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gszfv\" (UniqueName: \"kubernetes.io/projected/c30a71d3-ad60-465b-88f2-e642c701f598-kube-api-access-gszfv\") pod \"c30a71d3-ad60-465b-88f2-e642c701f598\" (UID: \"c30a71d3-ad60-465b-88f2-e642c701f598\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.601603 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30a71d3-ad60-465b-88f2-e642c701f598-operator-scripts\") pod \"c30a71d3-ad60-465b-88f2-e642c701f598\" (UID: \"c30a71d3-ad60-465b-88f2-e642c701f598\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.601671 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55hnn\" (UniqueName: \"kubernetes.io/projected/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-kube-api-access-55hnn\") pod \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\" (UID: \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.601738 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwdgj\" (UniqueName: \"kubernetes.io/projected/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-kube-api-access-dwdgj\") pod \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\" (UID: \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.601930 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-operator-scripts\") pod \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\" (UID: \"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.601982 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-operator-scripts\") pod \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\" (UID: \"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3\") " Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.602431 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3" (UID: "ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.602511 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c30a71d3-ad60-465b-88f2-e642c701f598-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c30a71d3-ad60-465b-88f2-e642c701f598" (UID: "c30a71d3-ad60-465b-88f2-e642c701f598"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.602951 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7fdf3ca-8982-4da7-8b15-1647ddb2fef3" (UID: "d7fdf3ca-8982-4da7-8b15-1647ddb2fef3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.603646 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsjnp\" (UniqueName: \"kubernetes.io/projected/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6-kube-api-access-nsjnp\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.603685 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30a71d3-ad60-465b-88f2-e642c701f598-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.603704 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj87v\" (UniqueName: \"kubernetes.io/projected/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3-kube-api-access-cj87v\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.603723 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.603740 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.605599 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-kube-api-access-55hnn" (OuterVolumeSpecName: "kube-api-access-55hnn") pod "ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3" (UID: "ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3"). InnerVolumeSpecName "kube-api-access-55hnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.606062 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c30a71d3-ad60-465b-88f2-e642c701f598-kube-api-access-gszfv" (OuterVolumeSpecName: "kube-api-access-gszfv") pod "c30a71d3-ad60-465b-88f2-e642c701f598" (UID: "c30a71d3-ad60-465b-88f2-e642c701f598"). InnerVolumeSpecName "kube-api-access-gszfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.607686 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-kube-api-access-dwdgj" (OuterVolumeSpecName: "kube-api-access-dwdgj") pod "d7fdf3ca-8982-4da7-8b15-1647ddb2fef3" (UID: "d7fdf3ca-8982-4da7-8b15-1647ddb2fef3"). InnerVolumeSpecName "kube-api-access-dwdgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.704947 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gszfv\" (UniqueName: \"kubernetes.io/projected/c30a71d3-ad60-465b-88f2-e642c701f598-kube-api-access-gszfv\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.705014 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55hnn\" (UniqueName: \"kubernetes.io/projected/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3-kube-api-access-55hnn\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.705024 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwdgj\" (UniqueName: \"kubernetes.io/projected/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3-kube-api-access-dwdgj\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.791398 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5vq76" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.791419 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5vq76" event={"ID":"40ea0fb9-d583-4ea8-8bb4-691e0f8606d3","Type":"ContainerDied","Data":"4ab907e76e224d156769209cdd2c3ffb5091c69dff5e6144c36da15b2900e76e"} Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.791482 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ab907e76e224d156769209cdd2c3ffb5091c69dff5e6144c36da15b2900e76e" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.793338 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-17fd-account-create-update-ct9z4" event={"ID":"c30a71d3-ad60-465b-88f2-e642c701f598","Type":"ContainerDied","Data":"aeff95349210fc793d2e056be60182b99092a0ad9242487190cdd006ec769268"} Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.793387 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aeff95349210fc793d2e056be60182b99092a0ad9242487190cdd006ec769268" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.793454 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-17fd-account-create-update-ct9z4" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.794907 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-76hpp" event={"ID":"41470d79-78b1-44bb-a10f-9f169d2d9e17","Type":"ContainerDied","Data":"bd56e6391be736cabdb0954cfab41f5d17c639bb1a0e07346bac10bbc89d15c4"} Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.794932 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd56e6391be736cabdb0954cfab41f5d17c639bb1a0e07346bac10bbc89d15c4" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.795050 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-76hpp" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.796425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2aa9-account-create-update-dd8zp" event={"ID":"ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3","Type":"ContainerDied","Data":"2b26d0391fe204d8a747bc3820108c9a015e516ccd97473b1229e6efe64d58b4"} Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.796449 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2aa9-account-create-update-dd8zp" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.796452 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b26d0391fe204d8a747bc3820108c9a015e516ccd97473b1229e6efe64d58b4" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.798678 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-q4prp" event={"ID":"d7fdf3ca-8982-4da7-8b15-1647ddb2fef3","Type":"ContainerDied","Data":"4ac312153d05e66c86ba777292d474b2e463be281bd0aac112fda771aedcb917"} Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.798726 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ac312153d05e66c86ba777292d474b2e463be281bd0aac112fda771aedcb917" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.800163 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-q4prp" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.801323 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7119-account-create-update-k2cvp" event={"ID":"0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6","Type":"ContainerDied","Data":"1a6565899dd98f4da2fcbc0839fea8104b651464cc611a59289961e5f1946a43"} Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.801363 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a6565899dd98f4da2fcbc0839fea8104b651464cc611a59289961e5f1946a43" Dec 04 17:59:40 crc kubenswrapper[4733]: I1204 17:59:40.801415 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7119-account-create-update-k2cvp" Dec 04 17:59:43 crc kubenswrapper[4733]: I1204 17:59:43.295837 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 17:59:43 crc kubenswrapper[4733]: I1204 17:59:43.361453 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-9b9k2"] Dec 04 17:59:43 crc kubenswrapper[4733]: I1204 17:59:43.366080 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" podUID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" containerName="dnsmasq-dns" containerID="cri-o://42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435" gracePeriod=10 Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.770694 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.836160 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-27f6z" event={"ID":"37aa673f-32a7-43ce-b1b2-a7f02f46c485","Type":"ContainerStarted","Data":"569e02b361ce9612c6faa846d0e150a84d2d650950ed49b4e0e824acc633aad4"} Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.840280 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.840371 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" event={"ID":"51afaec5-8ac0-47ab-a960-63bb01cf4fc1","Type":"ContainerDied","Data":"42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435"} Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.840404 4733 scope.go:117] "RemoveContainer" containerID="42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.840237 4733 generic.go:334] "Generic (PLEG): container finished" podID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" containerID="42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435" exitCode=0 Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.840522 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5fbdd8c-9b9k2" event={"ID":"51afaec5-8ac0-47ab-a960-63bb01cf4fc1","Type":"ContainerDied","Data":"0ff7a5bd86ca0753d79b844ffabd02f884b527ac493b22c4142ef691f3bf1072"} Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.858998 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-27f6z" podStartSLOduration=1.714237946 podStartE2EDuration="6.858981347s" podCreationTimestamp="2025-12-04 17:59:37 +0000 UTC" firstStartedPulling="2025-12-04 17:59:38.074644847 +0000 UTC m=+1240.030005893" lastFinishedPulling="2025-12-04 17:59:43.219388248 +0000 UTC m=+1245.174749294" observedRunningTime="2025-12-04 17:59:43.854659113 +0000 UTC m=+1245.810020159" watchObservedRunningTime="2025-12-04 17:59:43.858981347 +0000 UTC m=+1245.814342393" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.863755 4733 scope.go:117] "RemoveContainer" containerID="250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.884063 4733 scope.go:117] "RemoveContainer" containerID="42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435" Dec 04 17:59:44 crc kubenswrapper[4733]: E1204 17:59:43.884397 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435\": container with ID starting with 42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435 not found: ID does not exist" containerID="42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.884423 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435"} err="failed to get container status \"42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435\": rpc error: code = NotFound desc = could not find container \"42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435\": container with ID starting with 42b94308fe9f891628b8cae464edf8f3ac64f770f8f1889a58a56a2d9bd2d435 not found: ID does not exist" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.884442 4733 scope.go:117] "RemoveContainer" containerID="250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12" Dec 04 17:59:44 crc kubenswrapper[4733]: E1204 17:59:43.884709 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12\": container with ID starting with 250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12 not found: ID does not exist" containerID="250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.884724 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12"} err="failed to get container status \"250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12\": rpc error: code = NotFound desc = could not find container \"250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12\": container with ID starting with 250411ece72bc464ecf745f65c85d2d5890feb6381776dd86b4e78b737ffcc12 not found: ID does not exist" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.966632 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nptzp\" (UniqueName: \"kubernetes.io/projected/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-kube-api-access-nptzp\") pod \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.966821 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-sb\") pod \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.966911 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-config\") pod \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.966965 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-dns-svc\") pod \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.967012 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-nb\") pod \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\" (UID: \"51afaec5-8ac0-47ab-a960-63bb01cf4fc1\") " Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:43.972869 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-kube-api-access-nptzp" (OuterVolumeSpecName: "kube-api-access-nptzp") pod "51afaec5-8ac0-47ab-a960-63bb01cf4fc1" (UID: "51afaec5-8ac0-47ab-a960-63bb01cf4fc1"). InnerVolumeSpecName "kube-api-access-nptzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.008109 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "51afaec5-8ac0-47ab-a960-63bb01cf4fc1" (UID: "51afaec5-8ac0-47ab-a960-63bb01cf4fc1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.012145 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "51afaec5-8ac0-47ab-a960-63bb01cf4fc1" (UID: "51afaec5-8ac0-47ab-a960-63bb01cf4fc1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.022393 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51afaec5-8ac0-47ab-a960-63bb01cf4fc1" (UID: "51afaec5-8ac0-47ab-a960-63bb01cf4fc1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.034123 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-config" (OuterVolumeSpecName: "config") pod "51afaec5-8ac0-47ab-a960-63bb01cf4fc1" (UID: "51afaec5-8ac0-47ab-a960-63bb01cf4fc1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.069442 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.069482 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.069543 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.069563 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.069620 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nptzp\" (UniqueName: \"kubernetes.io/projected/51afaec5-8ac0-47ab-a960-63bb01cf4fc1-kube-api-access-nptzp\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.188946 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-9b9k2"] Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.199404 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5fbdd8c-9b9k2"] Dec 04 17:59:44 crc kubenswrapper[4733]: I1204 17:59:44.352891 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" path="/var/lib/kubelet/pods/51afaec5-8ac0-47ab-a960-63bb01cf4fc1/volumes" Dec 04 17:59:46 crc kubenswrapper[4733]: I1204 17:59:46.882000 4733 generic.go:334] "Generic (PLEG): container finished" podID="37aa673f-32a7-43ce-b1b2-a7f02f46c485" containerID="569e02b361ce9612c6faa846d0e150a84d2d650950ed49b4e0e824acc633aad4" exitCode=0 Dec 04 17:59:46 crc kubenswrapper[4733]: I1204 17:59:46.882177 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-27f6z" event={"ID":"37aa673f-32a7-43ce-b1b2-a7f02f46c485","Type":"ContainerDied","Data":"569e02b361ce9612c6faa846d0e150a84d2d650950ed49b4e0e824acc633aad4"} Dec 04 17:59:47 crc kubenswrapper[4733]: I1204 17:59:47.896729 4733 generic.go:334] "Generic (PLEG): container finished" podID="857cacd9-91d0-4a3a-92b8-73669fd75dfc" containerID="7daea43ad01541be1c71ecfc6cbf37ab9bcaf5bedff0f422646929d973829b99" exitCode=0 Dec 04 17:59:47 crc kubenswrapper[4733]: I1204 17:59:47.896938 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-p2kxj" event={"ID":"857cacd9-91d0-4a3a-92b8-73669fd75dfc","Type":"ContainerDied","Data":"7daea43ad01541be1c71ecfc6cbf37ab9bcaf5bedff0f422646929d973829b99"} Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.282489 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.447650 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jslsk\" (UniqueName: \"kubernetes.io/projected/37aa673f-32a7-43ce-b1b2-a7f02f46c485-kube-api-access-jslsk\") pod \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.448526 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-config-data\") pod \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.448626 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-combined-ca-bundle\") pod \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\" (UID: \"37aa673f-32a7-43ce-b1b2-a7f02f46c485\") " Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.454151 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37aa673f-32a7-43ce-b1b2-a7f02f46c485-kube-api-access-jslsk" (OuterVolumeSpecName: "kube-api-access-jslsk") pod "37aa673f-32a7-43ce-b1b2-a7f02f46c485" (UID: "37aa673f-32a7-43ce-b1b2-a7f02f46c485"). InnerVolumeSpecName "kube-api-access-jslsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.491328 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37aa673f-32a7-43ce-b1b2-a7f02f46c485" (UID: "37aa673f-32a7-43ce-b1b2-a7f02f46c485"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.510950 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-config-data" (OuterVolumeSpecName: "config-data") pod "37aa673f-32a7-43ce-b1b2-a7f02f46c485" (UID: "37aa673f-32a7-43ce-b1b2-a7f02f46c485"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.550043 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jslsk\" (UniqueName: \"kubernetes.io/projected/37aa673f-32a7-43ce-b1b2-a7f02f46c485-kube-api-access-jslsk\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.550083 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.550107 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37aa673f-32a7-43ce-b1b2-a7f02f46c485-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.912592 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-27f6z" Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.912668 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-27f6z" event={"ID":"37aa673f-32a7-43ce-b1b2-a7f02f46c485","Type":"ContainerDied","Data":"ba9e9e214d9e6c62c29b3b6c9ba0a43ad21db01046d3220beb84eb447fa7b420"} Dec 04 17:59:48 crc kubenswrapper[4733]: I1204 17:59:48.912704 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba9e9e214d9e6c62c29b3b6c9ba0a43ad21db01046d3220beb84eb447fa7b420" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.171687 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-4t7z6"] Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172156 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ea0fb9-d583-4ea8-8bb4-691e0f8606d3" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172174 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ea0fb9-d583-4ea8-8bb4-691e0f8606d3" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172189 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172196 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172214 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" containerName="dnsmasq-dns" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172222 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" containerName="dnsmasq-dns" Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172233 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41470d79-78b1-44bb-a10f-9f169d2d9e17" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172241 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="41470d79-78b1-44bb-a10f-9f169d2d9e17" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172257 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37aa673f-32a7-43ce-b1b2-a7f02f46c485" containerName="keystone-db-sync" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172264 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="37aa673f-32a7-43ce-b1b2-a7f02f46c485" containerName="keystone-db-sync" Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172287 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" containerName="init" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172296 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" containerName="init" Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172306 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7fdf3ca-8982-4da7-8b15-1647ddb2fef3" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172313 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7fdf3ca-8982-4da7-8b15-1647ddb2fef3" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172327 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30a71d3-ad60-465b-88f2-e642c701f598" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172334 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30a71d3-ad60-465b-88f2-e642c701f598" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.172348 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172355 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172544 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="51afaec5-8ac0-47ab-a960-63bb01cf4fc1" containerName="dnsmasq-dns" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172560 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c30a71d3-ad60-465b-88f2-e642c701f598" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172574 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172584 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3" containerName="mariadb-account-create-update" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172599 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="37aa673f-32a7-43ce-b1b2-a7f02f46c485" containerName="keystone-db-sync" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172613 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="40ea0fb9-d583-4ea8-8bb4-691e0f8606d3" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172626 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7fdf3ca-8982-4da7-8b15-1647ddb2fef3" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.172740 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="41470d79-78b1-44bb-a10f-9f169d2d9e17" containerName="mariadb-database-create" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.173818 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.191164 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-4t7z6"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.219646 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-cchrd"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.221069 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.224045 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.224259 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.224449 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r9kr4" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.224648 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.224762 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.238817 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cchrd"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.261746 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-scripts\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263133 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-svc\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263199 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-swift-storage-0\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263226 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-combined-ca-bundle\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263246 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-config-data\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263261 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-nb\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263286 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-credential-keys\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263334 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-fernet-keys\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263369 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkhpl\" (UniqueName: \"kubernetes.io/projected/235e2af3-b492-442c-9a9f-216f1bd1efed-kube-api-access-kkhpl\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263388 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-sb\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263420 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-config\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.263448 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjj76\" (UniqueName: \"kubernetes.io/projected/32af2fdb-0ded-4432-b2c5-7131630e5673-kube-api-access-jjj76\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366594 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-scripts\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366641 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-svc\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366664 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-swift-storage-0\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366687 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-combined-ca-bundle\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366706 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-config-data\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366722 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-nb\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366743 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-credential-keys\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366780 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-fernet-keys\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366821 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkhpl\" (UniqueName: \"kubernetes.io/projected/235e2af3-b492-442c-9a9f-216f1bd1efed-kube-api-access-kkhpl\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366841 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-sb\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366868 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-config\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.366897 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjj76\" (UniqueName: \"kubernetes.io/projected/32af2fdb-0ded-4432-b2c5-7131630e5673-kube-api-access-jjj76\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.369774 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-svc\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.369967 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-nb\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.371425 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-swift-storage-0\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.372278 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-sb\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.372362 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-config\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.388967 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjj76\" (UniqueName: \"kubernetes.io/projected/32af2fdb-0ded-4432-b2c5-7131630e5673-kube-api-access-jjj76\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.392546 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-combined-ca-bundle\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.392608 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2thjt"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.393581 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.394509 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-credential-keys\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.397492 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-config-data\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.403412 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.403667 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.403906 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rhhsz" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.410388 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkhpl\" (UniqueName: \"kubernetes.io/projected/235e2af3-b492-442c-9a9f-216f1bd1efed-kube-api-access-kkhpl\") pod \"dnsmasq-dns-78b9b4595f-4t7z6\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.410886 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-fernet-keys\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.411189 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-scripts\") pod \"keystone-bootstrap-cchrd\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.428992 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-btknr"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.430069 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.436583 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.436754 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.436963 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kh24l" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.468454 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2thjt"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.489033 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-btknr"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.513135 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.513564 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.516056 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.523871 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.524127 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.526508 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-gbxkg"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.529014 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.535744 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.535930 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jj8t9" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.536081 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.538574 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cchrd" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.543712 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gbxkg"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.570871 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-combined-ca-bundle\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.570929 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzx8t\" (UniqueName: \"kubernetes.io/projected/77bd0d5e-6bf1-4257-b30e-795688595c1c-kube-api-access-tzx8t\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.570947 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-scripts\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.570968 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bcc7f96-8d09-4d23-8e53-02b679d7851a-etc-machine-id\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.570989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-db-sync-config-data\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.571040 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkxzz\" (UniqueName: \"kubernetes.io/projected/3bcc7f96-8d09-4d23-8e53-02b679d7851a-kube-api-access-dkxzz\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.571060 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-combined-ca-bundle\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.571079 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-config-data\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.571100 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-config\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.631682 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-4t7z6"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.652487 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.657830 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-2khcb"] Dec 04 17:59:49 crc kubenswrapper[4733]: E1204 17:59:49.658179 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="857cacd9-91d0-4a3a-92b8-73669fd75dfc" containerName="glance-db-sync" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.658195 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="857cacd9-91d0-4a3a-92b8-73669fd75dfc" containerName="glance-db-sync" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.658396 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="857cacd9-91d0-4a3a-92b8-73669fd75dfc" containerName="glance-db-sync" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.659066 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.668454 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2khcb"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672534 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-combined-ca-bundle\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672604 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzx8t\" (UniqueName: \"kubernetes.io/projected/77bd0d5e-6bf1-4257-b30e-795688595c1c-kube-api-access-tzx8t\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672627 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-scripts\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672655 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88ssc\" (UniqueName: \"kubernetes.io/projected/abc26201-5065-4a19-9311-fc90d11401d0-kube-api-access-88ssc\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672678 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bcc7f96-8d09-4d23-8e53-02b679d7851a-etc-machine-id\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672700 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-scripts\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672715 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-config-data\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672732 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-db-sync-config-data\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672757 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-db-sync-config-data\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672778 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-log-httpd\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672823 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-run-httpd\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672857 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672930 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkxzz\" (UniqueName: \"kubernetes.io/projected/3bcc7f96-8d09-4d23-8e53-02b679d7851a-kube-api-access-dkxzz\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.672952 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9spkp\" (UniqueName: \"kubernetes.io/projected/3be601db-3698-4272-8005-083d5b93cbfd-kube-api-access-9spkp\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.673787 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bcc7f96-8d09-4d23-8e53-02b679d7851a-etc-machine-id\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.673965 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-combined-ca-bundle\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.674031 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-config-data\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.674070 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-config\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.674113 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.674136 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-combined-ca-bundle\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.679890 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-fnf9k"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.682160 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.685990 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.686180 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vw67t" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.686276 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.686897 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-config\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.703887 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-combined-ca-bundle\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.704592 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-scripts\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.717149 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-db-sync-config-data\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.729027 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzx8t\" (UniqueName: \"kubernetes.io/projected/77bd0d5e-6bf1-4257-b30e-795688595c1c-kube-api-access-tzx8t\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.730230 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-combined-ca-bundle\") pod \"neutron-db-sync-2thjt\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.730314 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-fnf9k"] Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.735253 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-config-data\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.741520 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkxzz\" (UniqueName: \"kubernetes.io/projected/3bcc7f96-8d09-4d23-8e53-02b679d7851a-kube-api-access-dkxzz\") pod \"cinder-db-sync-btknr\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.777814 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5nfd\" (UniqueName: \"kubernetes.io/projected/857cacd9-91d0-4a3a-92b8-73669fd75dfc-kube-api-access-n5nfd\") pod \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.777877 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-config-data\") pod \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.777991 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-db-sync-config-data\") pod \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778031 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-combined-ca-bundle\") pod \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\" (UID: \"857cacd9-91d0-4a3a-92b8-73669fd75dfc\") " Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778195 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778215 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-combined-ca-bundle\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778234 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-config-data\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778294 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-scripts\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778330 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88ssc\" (UniqueName: \"kubernetes.io/projected/abc26201-5065-4a19-9311-fc90d11401d0-kube-api-access-88ssc\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778356 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-scripts\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778373 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-config-data\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778391 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53e91d-1256-408a-9b97-a048f425c441-logs\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778407 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-db-sync-config-data\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778426 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-log-httpd\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778453 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-run-httpd\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778477 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778500 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9spkp\" (UniqueName: \"kubernetes.io/projected/3be601db-3698-4272-8005-083d5b93cbfd-kube-api-access-9spkp\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778523 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-combined-ca-bundle\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.778549 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krgsx\" (UniqueName: \"kubernetes.io/projected/8c53e91d-1256-408a-9b97-a048f425c441-kube-api-access-krgsx\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.782154 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/857cacd9-91d0-4a3a-92b8-73669fd75dfc-kube-api-access-n5nfd" (OuterVolumeSpecName: "kube-api-access-n5nfd") pod "857cacd9-91d0-4a3a-92b8-73669fd75dfc" (UID: "857cacd9-91d0-4a3a-92b8-73669fd75dfc"). InnerVolumeSpecName "kube-api-access-n5nfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.786043 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-scripts\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.790055 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-run-httpd\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.793677 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-log-httpd\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.797521 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.800267 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.800772 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-combined-ca-bundle\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.803825 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-db-sync-config-data\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.810558 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "857cacd9-91d0-4a3a-92b8-73669fd75dfc" (UID: "857cacd9-91d0-4a3a-92b8-73669fd75dfc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.814402 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-config-data\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.815027 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9spkp\" (UniqueName: \"kubernetes.io/projected/3be601db-3698-4272-8005-083d5b93cbfd-kube-api-access-9spkp\") pod \"barbican-db-sync-gbxkg\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.815113 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88ssc\" (UniqueName: \"kubernetes.io/projected/abc26201-5065-4a19-9311-fc90d11401d0-kube-api-access-88ssc\") pod \"ceilometer-0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.833201 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2thjt" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.840088 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "857cacd9-91d0-4a3a-92b8-73669fd75dfc" (UID: "857cacd9-91d0-4a3a-92b8-73669fd75dfc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.880574 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-scripts\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.880633 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-config\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.880651 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53e91d-1256-408a-9b97-a048f425c441-logs\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.880673 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-swift-storage-0\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.880708 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-sb\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.880978 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-combined-ca-bundle\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.881059 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53e91d-1256-408a-9b97-a048f425c441-logs\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.881544 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krgsx\" (UniqueName: \"kubernetes.io/projected/8c53e91d-1256-408a-9b97-a048f425c441-kube-api-access-krgsx\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.881608 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-nb\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.881677 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-config-data\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.881722 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-svc\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.882481 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbm6g\" (UniqueName: \"kubernetes.io/projected/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-kube-api-access-rbm6g\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.882681 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5nfd\" (UniqueName: \"kubernetes.io/projected/857cacd9-91d0-4a3a-92b8-73669fd75dfc-kube-api-access-n5nfd\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.882700 4733 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.882711 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.883712 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-scripts\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.884809 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-combined-ca-bundle\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.885461 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-config-data\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.890505 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-config-data" (OuterVolumeSpecName: "config-data") pod "857cacd9-91d0-4a3a-92b8-73669fd75dfc" (UID: "857cacd9-91d0-4a3a-92b8-73669fd75dfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.903642 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krgsx\" (UniqueName: \"kubernetes.io/projected/8c53e91d-1256-408a-9b97-a048f425c441-kube-api-access-krgsx\") pod \"placement-db-sync-2khcb\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.908489 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-btknr" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.925003 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-p2kxj" event={"ID":"857cacd9-91d0-4a3a-92b8-73669fd75dfc","Type":"ContainerDied","Data":"e389e6b35e9b6f5026b9684947719ccb287706e0c6166b6f991e182f5960a3e8"} Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.925047 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e389e6b35e9b6f5026b9684947719ccb287706e0c6166b6f991e182f5960a3e8" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.925114 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-p2kxj" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.947112 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.965966 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gbxkg" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.986726 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-nb\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.986898 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-svc\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.986981 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbm6g\" (UniqueName: \"kubernetes.io/projected/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-kube-api-access-rbm6g\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.987138 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-config\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.987176 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-swift-storage-0\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.987265 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-sb\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.987383 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857cacd9-91d0-4a3a-92b8-73669fd75dfc-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.987637 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-svc\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.988359 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-config\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.988558 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-sb\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.988688 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-swift-storage-0\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.988951 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2khcb" Dec 04 17:59:49 crc kubenswrapper[4733]: I1204 17:59:49.989786 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-nb\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.007364 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbm6g\" (UniqueName: \"kubernetes.io/projected/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-kube-api-access-rbm6g\") pod \"dnsmasq-dns-794c6877f7-fnf9k\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.047838 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.130239 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cchrd"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.166105 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-4t7z6"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.385010 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-fnf9k"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.465014 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-lpxxd"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.466625 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.506920 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2thjt"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.552767 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-lpxxd"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.585993 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-btknr"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.612098 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-config\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.612209 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.612308 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.612333 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-svc\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.612393 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzfrt\" (UniqueName: \"kubernetes.io/projected/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-kube-api-access-mzfrt\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.612454 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.717202 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzfrt\" (UniqueName: \"kubernetes.io/projected/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-kube-api-access-mzfrt\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.717269 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.717305 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-config\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.717349 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.717395 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.717416 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-svc\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.718345 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-svc\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.718416 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-config\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.719083 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-swift-storage-0\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.719428 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-nb\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.720070 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-sb\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.723485 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-fnf9k"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.743863 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2khcb"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.750701 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzfrt\" (UniqueName: \"kubernetes.io/projected/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-kube-api-access-mzfrt\") pod \"dnsmasq-dns-7884648fd9-lpxxd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.756168 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 17:59:50 crc kubenswrapper[4733]: W1204 17:59:50.780553 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1e6c3d6_0a8b_4077_81fd_48b249abf48b.slice/crio-57a17e26264f89c80e54e91435a4330dc5ab6080ba0150f5e3f62fb9fd5b09ff WatchSource:0}: Error finding container 57a17e26264f89c80e54e91435a4330dc5ab6080ba0150f5e3f62fb9fd5b09ff: Status 404 returned error can't find the container with id 57a17e26264f89c80e54e91435a4330dc5ab6080ba0150f5e3f62fb9fd5b09ff Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.909252 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gbxkg"] Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.959709 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:50 crc kubenswrapper[4733]: I1204 17:59:50.988084 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2khcb" event={"ID":"8c53e91d-1256-408a-9b97-a048f425c441","Type":"ContainerStarted","Data":"b860c02be4704ea5232d214ba1d2e57c2c5fe25dff5df1b04e653a3d7f852572"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.010898 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2thjt" event={"ID":"77bd0d5e-6bf1-4257-b30e-795688595c1c","Type":"ContainerStarted","Data":"d49942198858476882af324869180b206cbbcac041cbb1d3b56286f2de3f2d02"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.010950 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2thjt" event={"ID":"77bd0d5e-6bf1-4257-b30e-795688595c1c","Type":"ContainerStarted","Data":"670ab638276cc50e23ac089ddd4b8e12b115ef64fd8fa0dbef395a8460dc8f39"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.019941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-btknr" event={"ID":"3bcc7f96-8d09-4d23-8e53-02b679d7851a","Type":"ContainerStarted","Data":"fdbe17d093c9b8fe804454f2ab16a6d587d3fc47e85a0be922588a09409647cf"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.035032 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" event={"ID":"b1e6c3d6-0a8b-4077-81fd-48b249abf48b","Type":"ContainerStarted","Data":"57a17e26264f89c80e54e91435a4330dc5ab6080ba0150f5e3f62fb9fd5b09ff"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.036670 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"abc26201-5065-4a19-9311-fc90d11401d0","Type":"ContainerStarted","Data":"1dc2d491ae83567fd2aef13dccfba33cd1ee64e757d0786ae9951e67fd0e96f0"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.039294 4733 generic.go:334] "Generic (PLEG): container finished" podID="235e2af3-b492-442c-9a9f-216f1bd1efed" containerID="f04a82a3646714f668a4708c5ae2106e8db97120cd483f2dc034126f6ed0b76a" exitCode=0 Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.039342 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" event={"ID":"235e2af3-b492-442c-9a9f-216f1bd1efed","Type":"ContainerDied","Data":"f04a82a3646714f668a4708c5ae2106e8db97120cd483f2dc034126f6ed0b76a"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.039360 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" event={"ID":"235e2af3-b492-442c-9a9f-216f1bd1efed","Type":"ContainerStarted","Data":"60526d2659c2c60f6c7e0fe5f8bca758478d593aafaeee8713d051a84a6bd6d4"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.042912 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2thjt" podStartSLOduration=2.04290131 podStartE2EDuration="2.04290131s" podCreationTimestamp="2025-12-04 17:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:59:51.032042375 +0000 UTC m=+1252.987403491" watchObservedRunningTime="2025-12-04 17:59:51.04290131 +0000 UTC m=+1252.998262356" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.073286 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cchrd" event={"ID":"32af2fdb-0ded-4432-b2c5-7131630e5673","Type":"ContainerStarted","Data":"2335b1cf12fa0b82828e2f0e53c505bbc80375fc5b0a4726aee29ed9a4847793"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.073326 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cchrd" event={"ID":"32af2fdb-0ded-4432-b2c5-7131630e5673","Type":"ContainerStarted","Data":"24ffd7047484b3b2bf3bb87d7b07d41a38e5d1284362fd3d9a70650d461f9e30"} Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.111460 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-cchrd" podStartSLOduration=2.111443093 podStartE2EDuration="2.111443093s" podCreationTimestamp="2025-12-04 17:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:59:51.108877975 +0000 UTC m=+1253.064239041" watchObservedRunningTime="2025-12-04 17:59:51.111443093 +0000 UTC m=+1253.066804139" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.378244 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.381501 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.383515 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g46cp" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.384368 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.387275 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.409095 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.467980 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.480011 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.486569 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.489091 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.526359 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.542820 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-scripts\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.542874 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgmtr\" (UniqueName: \"kubernetes.io/projected/c2afdd33-69d5-4717-ac11-3182b542963d-kube-api-access-tgmtr\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.542896 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.542936 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.542963 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-config-data\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.542985 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.543019 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-logs\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: E1204 17:59:51.555834 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run kube-api-access-tgmtr logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-external-api-0" podUID="c2afdd33-69d5-4717-ac11-3182b542963d" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.562065 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.569660 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:51 crc kubenswrapper[4733]: E1204 17:59:51.577018 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run kube-api-access-pkp97 logs scripts], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data glance httpd-run kube-api-access-pkp97 logs scripts]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="3f561b91-4f96-420a-b212-fb5c355a531a" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.610496 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.645548 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkhpl\" (UniqueName: \"kubernetes.io/projected/235e2af3-b492-442c-9a9f-216f1bd1efed-kube-api-access-kkhpl\") pod \"235e2af3-b492-442c-9a9f-216f1bd1efed\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.645607 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-sb\") pod \"235e2af3-b492-442c-9a9f-216f1bd1efed\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.645659 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-svc\") pod \"235e2af3-b492-442c-9a9f-216f1bd1efed\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.645679 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-config\") pod \"235e2af3-b492-442c-9a9f-216f1bd1efed\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.645817 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-nb\") pod \"235e2af3-b492-442c-9a9f-216f1bd1efed\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.645865 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-swift-storage-0\") pod \"235e2af3-b492-442c-9a9f-216f1bd1efed\" (UID: \"235e2af3-b492-442c-9a9f-216f1bd1efed\") " Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646059 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-scripts\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646097 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646131 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgmtr\" (UniqueName: \"kubernetes.io/projected/c2afdd33-69d5-4717-ac11-3182b542963d-kube-api-access-tgmtr\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646151 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646167 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646207 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646229 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646254 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-config-data\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646276 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646313 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-logs\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646342 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkp97\" (UniqueName: \"kubernetes.io/projected/3f561b91-4f96-420a-b212-fb5c355a531a-kube-api-access-pkp97\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646364 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646380 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.646399 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.647606 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.650238 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-logs\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.652497 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.665464 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.675234 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235e2af3-b492-442c-9a9f-216f1bd1efed-kube-api-access-kkhpl" (OuterVolumeSpecName: "kube-api-access-kkhpl") pod "235e2af3-b492-442c-9a9f-216f1bd1efed" (UID: "235e2af3-b492-442c-9a9f-216f1bd1efed"). InnerVolumeSpecName "kube-api-access-kkhpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.681157 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-scripts\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.697031 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-config-data\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.702095 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgmtr\" (UniqueName: \"kubernetes.io/projected/c2afdd33-69d5-4717-ac11-3182b542963d-kube-api-access-tgmtr\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.709554 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.725551 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-lpxxd"] Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.726845 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-config" (OuterVolumeSpecName: "config") pod "235e2af3-b492-442c-9a9f-216f1bd1efed" (UID: "235e2af3-b492-442c-9a9f-216f1bd1efed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749554 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749598 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749624 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749661 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749688 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749730 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749811 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkp97\" (UniqueName: \"kubernetes.io/projected/3f561b91-4f96-420a-b212-fb5c355a531a-kube-api-access-pkp97\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749856 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkhpl\" (UniqueName: \"kubernetes.io/projected/235e2af3-b492-442c-9a9f-216f1bd1efed-kube-api-access-kkhpl\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.749872 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.750165 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.750315 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.750745 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.756151 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "235e2af3-b492-442c-9a9f-216f1bd1efed" (UID: "235e2af3-b492-442c-9a9f-216f1bd1efed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.762566 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "235e2af3-b492-442c-9a9f-216f1bd1efed" (UID: "235e2af3-b492-442c-9a9f-216f1bd1efed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:51 crc kubenswrapper[4733]: W1204 17:59:51.762653 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc20a96e3_9c22_48fc_ba00_4f1a13350ecd.slice/crio-8f264d3b1793a9e00a2e15c65efd8f18f68d8e57af3b40ae1dbd8c0e5793a658 WatchSource:0}: Error finding container 8f264d3b1793a9e00a2e15c65efd8f18f68d8e57af3b40ae1dbd8c0e5793a658: Status 404 returned error can't find the container with id 8f264d3b1793a9e00a2e15c65efd8f18f68d8e57af3b40ae1dbd8c0e5793a658 Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.762706 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.763142 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.763262 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.766460 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkp97\" (UniqueName: \"kubernetes.io/projected/3f561b91-4f96-420a-b212-fb5c355a531a-kube-api-access-pkp97\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.779303 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "235e2af3-b492-442c-9a9f-216f1bd1efed" (UID: "235e2af3-b492-442c-9a9f-216f1bd1efed"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.797203 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "235e2af3-b492-442c-9a9f-216f1bd1efed" (UID: "235e2af3-b492-442c-9a9f-216f1bd1efed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.801427 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.851261 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.851294 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.851303 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:51 crc kubenswrapper[4733]: I1204 17:59:51.851313 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/235e2af3-b492-442c-9a9f-216f1bd1efed-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.085919 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gbxkg" event={"ID":"3be601db-3698-4272-8005-083d5b93cbfd","Type":"ContainerStarted","Data":"0723534f5e69a040f6f2edc098bea25ecbb6be9543d7abd95820e56edb8a3af4"} Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.087693 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" event={"ID":"235e2af3-b492-442c-9a9f-216f1bd1efed","Type":"ContainerDied","Data":"60526d2659c2c60f6c7e0fe5f8bca758478d593aafaeee8713d051a84a6bd6d4"} Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.087739 4733 scope.go:117] "RemoveContainer" containerID="f04a82a3646714f668a4708c5ae2106e8db97120cd483f2dc034126f6ed0b76a" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.087748 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b9b4595f-4t7z6" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.095825 4733 generic.go:334] "Generic (PLEG): container finished" podID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" containerID="f5e6f2c72112a2171e2ad78efad46bcf8b2b83073efb9334f19fa80660ebe091" exitCode=0 Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.095879 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" event={"ID":"c20a96e3-9c22-48fc-ba00-4f1a13350ecd","Type":"ContainerDied","Data":"f5e6f2c72112a2171e2ad78efad46bcf8b2b83073efb9334f19fa80660ebe091"} Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.095905 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" event={"ID":"c20a96e3-9c22-48fc-ba00-4f1a13350ecd","Type":"ContainerStarted","Data":"8f264d3b1793a9e00a2e15c65efd8f18f68d8e57af3b40ae1dbd8c0e5793a658"} Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.100986 4733 generic.go:334] "Generic (PLEG): container finished" podID="b1e6c3d6-0a8b-4077-81fd-48b249abf48b" containerID="e110022cf4f041880c7932c8c3c33becac5671dc635db4546736075f68259669" exitCode=0 Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.101250 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.101963 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" event={"ID":"b1e6c3d6-0a8b-4077-81fd-48b249abf48b","Type":"ContainerDied","Data":"e110022cf4f041880c7932c8c3c33becac5671dc635db4546736075f68259669"} Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.102016 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.112228 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.154999 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-logs\") pod \"3f561b91-4f96-420a-b212-fb5c355a531a\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.155081 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-config-data\") pod \"3f561b91-4f96-420a-b212-fb5c355a531a\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.155115 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-combined-ca-bundle\") pod \"3f561b91-4f96-420a-b212-fb5c355a531a\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.155134 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-httpd-run\") pod \"3f561b91-4f96-420a-b212-fb5c355a531a\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.155162 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkp97\" (UniqueName: \"kubernetes.io/projected/3f561b91-4f96-420a-b212-fb5c355a531a-kube-api-access-pkp97\") pod \"3f561b91-4f96-420a-b212-fb5c355a531a\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.155421 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"3f561b91-4f96-420a-b212-fb5c355a531a\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.155558 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-scripts\") pod \"3f561b91-4f96-420a-b212-fb5c355a531a\" (UID: \"3f561b91-4f96-420a-b212-fb5c355a531a\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.156871 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-logs" (OuterVolumeSpecName: "logs") pod "3f561b91-4f96-420a-b212-fb5c355a531a" (UID: "3f561b91-4f96-420a-b212-fb5c355a531a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.162422 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-config-data" (OuterVolumeSpecName: "config-data") pod "3f561b91-4f96-420a-b212-fb5c355a531a" (UID: "3f561b91-4f96-420a-b212-fb5c355a531a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.164071 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f561b91-4f96-420a-b212-fb5c355a531a" (UID: "3f561b91-4f96-420a-b212-fb5c355a531a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.164338 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3f561b91-4f96-420a-b212-fb5c355a531a" (UID: "3f561b91-4f96-420a-b212-fb5c355a531a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.164915 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.168554 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f561b91-4f96-420a-b212-fb5c355a531a-kube-api-access-pkp97" (OuterVolumeSpecName: "kube-api-access-pkp97") pod "3f561b91-4f96-420a-b212-fb5c355a531a" (UID: "3f561b91-4f96-420a-b212-fb5c355a531a"). InnerVolumeSpecName "kube-api-access-pkp97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.176876 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-scripts" (OuterVolumeSpecName: "scripts") pod "3f561b91-4f96-420a-b212-fb5c355a531a" (UID: "3f561b91-4f96-420a-b212-fb5c355a531a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.189036 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "3f561b91-4f96-420a-b212-fb5c355a531a" (UID: "3f561b91-4f96-420a-b212-fb5c355a531a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.258424 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-combined-ca-bundle\") pod \"c2afdd33-69d5-4717-ac11-3182b542963d\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.258493 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-config-data\") pod \"c2afdd33-69d5-4717-ac11-3182b542963d\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.258585 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-scripts\") pod \"c2afdd33-69d5-4717-ac11-3182b542963d\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.258845 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgmtr\" (UniqueName: \"kubernetes.io/projected/c2afdd33-69d5-4717-ac11-3182b542963d-kube-api-access-tgmtr\") pod \"c2afdd33-69d5-4717-ac11-3182b542963d\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.258884 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-httpd-run\") pod \"c2afdd33-69d5-4717-ac11-3182b542963d\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.258910 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-logs\") pod \"c2afdd33-69d5-4717-ac11-3182b542963d\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.258931 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"c2afdd33-69d5-4717-ac11-3182b542963d\" (UID: \"c2afdd33-69d5-4717-ac11-3182b542963d\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.259408 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-logs\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.259419 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.259430 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.259444 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f561b91-4f96-420a-b212-fb5c355a531a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.259454 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkp97\" (UniqueName: \"kubernetes.io/projected/3f561b91-4f96-420a-b212-fb5c355a531a-kube-api-access-pkp97\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.259481 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.259496 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f561b91-4f96-420a-b212-fb5c355a531a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.260587 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-logs" (OuterVolumeSpecName: "logs") pod "c2afdd33-69d5-4717-ac11-3182b542963d" (UID: "c2afdd33-69d5-4717-ac11-3182b542963d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.262629 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c2afdd33-69d5-4717-ac11-3182b542963d" (UID: "c2afdd33-69d5-4717-ac11-3182b542963d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.262993 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-4t7z6"] Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.266519 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-scripts" (OuterVolumeSpecName: "scripts") pod "c2afdd33-69d5-4717-ac11-3182b542963d" (UID: "c2afdd33-69d5-4717-ac11-3182b542963d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.267924 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78b9b4595f-4t7z6"] Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.268879 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2afdd33-69d5-4717-ac11-3182b542963d" (UID: "c2afdd33-69d5-4717-ac11-3182b542963d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.269103 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2afdd33-69d5-4717-ac11-3182b542963d-kube-api-access-tgmtr" (OuterVolumeSpecName: "kube-api-access-tgmtr") pod "c2afdd33-69d5-4717-ac11-3182b542963d" (UID: "c2afdd33-69d5-4717-ac11-3182b542963d"). InnerVolumeSpecName "kube-api-access-tgmtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.274962 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-config-data" (OuterVolumeSpecName: "config-data") pod "c2afdd33-69d5-4717-ac11-3182b542963d" (UID: "c2afdd33-69d5-4717-ac11-3182b542963d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.276599 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "c2afdd33-69d5-4717-ac11-3182b542963d" (UID: "c2afdd33-69d5-4717-ac11-3182b542963d"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.333951 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.361471 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgmtr\" (UniqueName: \"kubernetes.io/projected/c2afdd33-69d5-4717-ac11-3182b542963d-kube-api-access-tgmtr\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.361783 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235e2af3-b492-442c-9a9f-216f1bd1efed" path="/var/lib/kubelet/pods/235e2af3-b492-442c-9a9f-216f1bd1efed/volumes" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.364234 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.381896 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.382007 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2afdd33-69d5-4717-ac11-3182b542963d-logs\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.382041 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.382066 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.382080 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.382089 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2afdd33-69d5-4717-ac11-3182b542963d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.425387 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.483775 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.743686 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.790449 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-sb\") pod \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.790524 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-config\") pod \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.790593 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-nb\") pod \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.790661 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-swift-storage-0\") pod \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.790688 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-svc\") pod \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.790726 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbm6g\" (UniqueName: \"kubernetes.io/projected/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-kube-api-access-rbm6g\") pod \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\" (UID: \"b1e6c3d6-0a8b-4077-81fd-48b249abf48b\") " Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.799015 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-kube-api-access-rbm6g" (OuterVolumeSpecName: "kube-api-access-rbm6g") pod "b1e6c3d6-0a8b-4077-81fd-48b249abf48b" (UID: "b1e6c3d6-0a8b-4077-81fd-48b249abf48b"). InnerVolumeSpecName "kube-api-access-rbm6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.816832 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b1e6c3d6-0a8b-4077-81fd-48b249abf48b" (UID: "b1e6c3d6-0a8b-4077-81fd-48b249abf48b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.827585 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b1e6c3d6-0a8b-4077-81fd-48b249abf48b" (UID: "b1e6c3d6-0a8b-4077-81fd-48b249abf48b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.837682 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b1e6c3d6-0a8b-4077-81fd-48b249abf48b" (UID: "b1e6c3d6-0a8b-4077-81fd-48b249abf48b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.838271 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b1e6c3d6-0a8b-4077-81fd-48b249abf48b" (UID: "b1e6c3d6-0a8b-4077-81fd-48b249abf48b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.869032 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-config" (OuterVolumeSpecName: "config") pod "b1e6c3d6-0a8b-4077-81fd-48b249abf48b" (UID: "b1e6c3d6-0a8b-4077-81fd-48b249abf48b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.892562 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.892593 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-config\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.892603 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.892612 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.892622 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:52 crc kubenswrapper[4733]: I1204 17:59:52.892630 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbm6g\" (UniqueName: \"kubernetes.io/projected/b1e6c3d6-0a8b-4077-81fd-48b249abf48b-kube-api-access-rbm6g\") on node \"crc\" DevicePath \"\"" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.114422 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" event={"ID":"b1e6c3d6-0a8b-4077-81fd-48b249abf48b","Type":"ContainerDied","Data":"57a17e26264f89c80e54e91435a4330dc5ab6080ba0150f5e3f62fb9fd5b09ff"} Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.114452 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794c6877f7-fnf9k" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.114473 4733 scope.go:117] "RemoveContainer" containerID="e110022cf4f041880c7932c8c3c33becac5671dc635db4546736075f68259669" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.116318 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.116319 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.221907 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.239388 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.244957 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 17:59:53 crc kubenswrapper[4733]: E1204 17:59:53.245381 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e6c3d6-0a8b-4077-81fd-48b249abf48b" containerName="init" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.245395 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e6c3d6-0a8b-4077-81fd-48b249abf48b" containerName="init" Dec 04 17:59:53 crc kubenswrapper[4733]: E1204 17:59:53.245417 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235e2af3-b492-442c-9a9f-216f1bd1efed" containerName="init" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.245423 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="235e2af3-b492-442c-9a9f-216f1bd1efed" containerName="init" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.245599 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="235e2af3-b492-442c-9a9f-216f1bd1efed" containerName="init" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.245615 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e6c3d6-0a8b-4077-81fd-48b249abf48b" containerName="init" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.246457 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.251586 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g46cp" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.251787 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.252292 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.264633 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.281110 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.291253 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.298936 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.298995 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.299031 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.299081 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-logs\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.299132 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.299158 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.299190 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vm97\" (UniqueName: \"kubernetes.io/projected/51031e1a-06fe-441e-a9d7-97523825a26e-kube-api-access-5vm97\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.299224 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.301592 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.307903 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.325764 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-fnf9k"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.331579 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-794c6877f7-fnf9k"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.337785 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.400871 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.400908 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc44h\" (UniqueName: \"kubernetes.io/projected/40314bda-46f4-426e-aeff-0029afdf5c90-kube-api-access-zc44h\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.400934 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-logs\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.400952 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-scripts\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401028 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401046 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-config-data\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401071 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-logs\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401128 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401145 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401168 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401188 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401213 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vm97\" (UniqueName: \"kubernetes.io/projected/51031e1a-06fe-441e-a9d7-97523825a26e-kube-api-access-5vm97\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401250 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401267 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401722 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.402776 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-logs\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.401825 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.408757 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.408830 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.409074 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.418447 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vm97\" (UniqueName: \"kubernetes.io/projected/51031e1a-06fe-441e-a9d7-97523825a26e-kube-api-access-5vm97\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.434326 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.505630 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc44h\" (UniqueName: \"kubernetes.io/projected/40314bda-46f4-426e-aeff-0029afdf5c90-kube-api-access-zc44h\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.505692 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-logs\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.505716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-scripts\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.505773 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.505813 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-config-data\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.505863 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.505894 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.506841 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.506848 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.507071 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-logs\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.510095 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-scripts\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.510398 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.511364 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-config-data\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.534299 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc44h\" (UniqueName: \"kubernetes.io/projected/40314bda-46f4-426e-aeff-0029afdf5c90-kube-api-access-zc44h\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.534674 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " pod="openstack/glance-default-external-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.580460 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 17:59:53 crc kubenswrapper[4733]: I1204 17:59:53.624480 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 17:59:54 crc kubenswrapper[4733]: I1204 17:59:54.352385 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f561b91-4f96-420a-b212-fb5c355a531a" path="/var/lib/kubelet/pods/3f561b91-4f96-420a-b212-fb5c355a531a/volumes" Dec 04 17:59:54 crc kubenswrapper[4733]: I1204 17:59:54.353339 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e6c3d6-0a8b-4077-81fd-48b249abf48b" path="/var/lib/kubelet/pods/b1e6c3d6-0a8b-4077-81fd-48b249abf48b/volumes" Dec 04 17:59:54 crc kubenswrapper[4733]: I1204 17:59:54.354429 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2afdd33-69d5-4717-ac11-3182b542963d" path="/var/lib/kubelet/pods/c2afdd33-69d5-4717-ac11-3182b542963d/volumes" Dec 04 17:59:55 crc kubenswrapper[4733]: I1204 17:59:55.332625 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 17:59:55 crc kubenswrapper[4733]: W1204 17:59:55.340344 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51031e1a_06fe_441e_a9d7_97523825a26e.slice/crio-7a6dff60f8934ffdbe65b066586d3087c7fdc86a69c97b91582b969f4bd58efc WatchSource:0}: Error finding container 7a6dff60f8934ffdbe65b066586d3087c7fdc86a69c97b91582b969f4bd58efc: Status 404 returned error can't find the container with id 7a6dff60f8934ffdbe65b066586d3087c7fdc86a69c97b91582b969f4bd58efc Dec 04 17:59:56 crc kubenswrapper[4733]: I1204 17:59:56.047813 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:56 crc kubenswrapper[4733]: I1204 17:59:56.153936 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51031e1a-06fe-441e-a9d7-97523825a26e","Type":"ContainerStarted","Data":"8db2bb8eb14546ebea7175ee3d8b0c49509e28fb3250d5aa493e06d670d5943c"} Dec 04 17:59:56 crc kubenswrapper[4733]: I1204 17:59:56.154307 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51031e1a-06fe-441e-a9d7-97523825a26e","Type":"ContainerStarted","Data":"7a6dff60f8934ffdbe65b066586d3087c7fdc86a69c97b91582b969f4bd58efc"} Dec 04 17:59:56 crc kubenswrapper[4733]: I1204 17:59:56.158332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" event={"ID":"c20a96e3-9c22-48fc-ba00-4f1a13350ecd","Type":"ContainerStarted","Data":"099a0dd66081b1742a35d04c3412573bb6ed9b1b5b14e8add80a8d0e79aa25ab"} Dec 04 17:59:56 crc kubenswrapper[4733]: I1204 17:59:56.158528 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 17:59:56 crc kubenswrapper[4733]: I1204 17:59:56.179613 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" podStartSLOduration=6.179596078 podStartE2EDuration="6.179596078s" podCreationTimestamp="2025-12-04 17:59:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 17:59:56.17700212 +0000 UTC m=+1258.132363186" watchObservedRunningTime="2025-12-04 17:59:56.179596078 +0000 UTC m=+1258.134957124" Dec 04 17:59:59 crc kubenswrapper[4733]: I1204 17:59:59.208863 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 17:59:59 crc kubenswrapper[4733]: I1204 17:59:59.276025 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.134251 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml"] Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.135370 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.138039 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.138727 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.142146 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml"] Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.196357 4733 generic.go:334] "Generic (PLEG): container finished" podID="32af2fdb-0ded-4432-b2c5-7131630e5673" containerID="2335b1cf12fa0b82828e2f0e53c505bbc80375fc5b0a4726aee29ed9a4847793" exitCode=0 Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.196396 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cchrd" event={"ID":"32af2fdb-0ded-4432-b2c5-7131630e5673","Type":"ContainerDied","Data":"2335b1cf12fa0b82828e2f0e53c505bbc80375fc5b0a4726aee29ed9a4847793"} Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.230041 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce93f96-1275-437f-9bcd-a7b299bae008-config-volume\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.230142 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce93f96-1275-437f-9bcd-a7b299bae008-secret-volume\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.230222 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zczw9\" (UniqueName: \"kubernetes.io/projected/4ce93f96-1275-437f-9bcd-a7b299bae008-kube-api-access-zczw9\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.333081 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce93f96-1275-437f-9bcd-a7b299bae008-secret-volume\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.333201 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zczw9\" (UniqueName: \"kubernetes.io/projected/4ce93f96-1275-437f-9bcd-a7b299bae008-kube-api-access-zczw9\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.333269 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce93f96-1275-437f-9bcd-a7b299bae008-config-volume\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.334451 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce93f96-1275-437f-9bcd-a7b299bae008-config-volume\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.357273 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zczw9\" (UniqueName: \"kubernetes.io/projected/4ce93f96-1275-437f-9bcd-a7b299bae008-kube-api-access-zczw9\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.361853 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce93f96-1275-437f-9bcd-a7b299bae008-secret-volume\") pod \"collect-profiles-29414520-nzxml\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.462920 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:00 crc kubenswrapper[4733]: I1204 18:00:00.962041 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 18:00:01 crc kubenswrapper[4733]: I1204 18:00:01.018103 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-ptkcg"] Dec 04 18:00:01 crc kubenswrapper[4733]: I1204 18:00:01.018326 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerName="dnsmasq-dns" containerID="cri-o://8d980a83191192ad31b4a99fa1e5f80399e038222d823f299b5b3eac1a730fda" gracePeriod=10 Dec 04 18:00:01 crc kubenswrapper[4733]: I1204 18:00:01.207569 4733 generic.go:334] "Generic (PLEG): container finished" podID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerID="8d980a83191192ad31b4a99fa1e5f80399e038222d823f299b5b3eac1a730fda" exitCode=0 Dec 04 18:00:01 crc kubenswrapper[4733]: I1204 18:00:01.207647 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" event={"ID":"5fdc02f1-40d7-43a1-9869-7815e0e6d2db","Type":"ContainerDied","Data":"8d980a83191192ad31b4a99fa1e5f80399e038222d823f299b5b3eac1a730fda"} Dec 04 18:00:01 crc kubenswrapper[4733]: I1204 18:00:01.208901 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"40314bda-46f4-426e-aeff-0029afdf5c90","Type":"ContainerStarted","Data":"7e20574e0ea3bc22f3ad1eff735250e2f9c5e96909cab5495e5a236a09860624"} Dec 04 18:00:03 crc kubenswrapper[4733]: I1204 18:00:03.294142 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.293902 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Dec 04 18:00:08 crc kubenswrapper[4733]: E1204 18:00:08.688278 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31" Dec 04 18:00:08 crc kubenswrapper[4733]: E1204 18:00:08.688479 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd7h5c9h577h5c5hd8h8fh696h5cfhbch679h5f8h546hfbh99h56hcfh75h657h89h677h584h59hffh579h5b5h688hc5hf6h684h97h9bhf9q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-88ssc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(abc26201-5065-4a19-9311-fc90d11401d0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.763282 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cchrd" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.788926 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjj76\" (UniqueName: \"kubernetes.io/projected/32af2fdb-0ded-4432-b2c5-7131630e5673-kube-api-access-jjj76\") pod \"32af2fdb-0ded-4432-b2c5-7131630e5673\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.789004 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-config-data\") pod \"32af2fdb-0ded-4432-b2c5-7131630e5673\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.789155 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-credential-keys\") pod \"32af2fdb-0ded-4432-b2c5-7131630e5673\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.789212 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-scripts\") pod \"32af2fdb-0ded-4432-b2c5-7131630e5673\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.789284 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-combined-ca-bundle\") pod \"32af2fdb-0ded-4432-b2c5-7131630e5673\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.789312 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-fernet-keys\") pod \"32af2fdb-0ded-4432-b2c5-7131630e5673\" (UID: \"32af2fdb-0ded-4432-b2c5-7131630e5673\") " Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.799099 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "32af2fdb-0ded-4432-b2c5-7131630e5673" (UID: "32af2fdb-0ded-4432-b2c5-7131630e5673"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.804155 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32af2fdb-0ded-4432-b2c5-7131630e5673-kube-api-access-jjj76" (OuterVolumeSpecName: "kube-api-access-jjj76") pod "32af2fdb-0ded-4432-b2c5-7131630e5673" (UID: "32af2fdb-0ded-4432-b2c5-7131630e5673"). InnerVolumeSpecName "kube-api-access-jjj76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.806847 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "32af2fdb-0ded-4432-b2c5-7131630e5673" (UID: "32af2fdb-0ded-4432-b2c5-7131630e5673"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.809250 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-scripts" (OuterVolumeSpecName: "scripts") pod "32af2fdb-0ded-4432-b2c5-7131630e5673" (UID: "32af2fdb-0ded-4432-b2c5-7131630e5673"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.834652 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32af2fdb-0ded-4432-b2c5-7131630e5673" (UID: "32af2fdb-0ded-4432-b2c5-7131630e5673"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.835753 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-config-data" (OuterVolumeSpecName: "config-data") pod "32af2fdb-0ded-4432-b2c5-7131630e5673" (UID: "32af2fdb-0ded-4432-b2c5-7131630e5673"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.890954 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.890991 4733 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.891002 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.891010 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.891019 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32af2fdb-0ded-4432-b2c5-7131630e5673-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:08 crc kubenswrapper[4733]: I1204 18:00:08.891028 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjj76\" (UniqueName: \"kubernetes.io/projected/32af2fdb-0ded-4432-b2c5-7131630e5673-kube-api-access-jjj76\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.297895 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cchrd" event={"ID":"32af2fdb-0ded-4432-b2c5-7131630e5673","Type":"ContainerDied","Data":"24ffd7047484b3b2bf3bb87d7b07d41a38e5d1284362fd3d9a70650d461f9e30"} Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.297946 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24ffd7047484b3b2bf3bb87d7b07d41a38e5d1284362fd3d9a70650d461f9e30" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.297979 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cchrd" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.855243 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-cchrd"] Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.863635 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-cchrd"] Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.956729 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nkhjz"] Dec 04 18:00:09 crc kubenswrapper[4733]: E1204 18:00:09.957311 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32af2fdb-0ded-4432-b2c5-7131630e5673" containerName="keystone-bootstrap" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.957337 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="32af2fdb-0ded-4432-b2c5-7131630e5673" containerName="keystone-bootstrap" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.957517 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="32af2fdb-0ded-4432-b2c5-7131630e5673" containerName="keystone-bootstrap" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.958586 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.960937 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r9kr4" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.961001 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.962321 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.962407 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.962486 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 18:00:09 crc kubenswrapper[4733]: I1204 18:00:09.966696 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nkhjz"] Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.006832 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-combined-ca-bundle\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.006890 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-credential-keys\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.006976 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-577s5\" (UniqueName: \"kubernetes.io/projected/2fca0e02-094d-4588-9efa-09181761155a-kube-api-access-577s5\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.007011 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-config-data\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.007053 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-fernet-keys\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.007073 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-scripts\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.108553 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-combined-ca-bundle\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.108601 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-credential-keys\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.108660 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-577s5\" (UniqueName: \"kubernetes.io/projected/2fca0e02-094d-4588-9efa-09181761155a-kube-api-access-577s5\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.108717 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-config-data\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.108758 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-scripts\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.108780 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-fernet-keys\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.113141 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-scripts\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.115263 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-config-data\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.116825 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-fernet-keys\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.117753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-combined-ca-bundle\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.132364 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-577s5\" (UniqueName: \"kubernetes.io/projected/2fca0e02-094d-4588-9efa-09181761155a-kube-api-access-577s5\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.132421 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-credential-keys\") pod \"keystone-bootstrap-nkhjz\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.279901 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:10 crc kubenswrapper[4733]: I1204 18:00:10.346833 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32af2fdb-0ded-4432-b2c5-7131630e5673" path="/var/lib/kubelet/pods/32af2fdb-0ded-4432-b2c5-7131630e5673/volumes" Dec 04 18:00:11 crc kubenswrapper[4733]: E1204 18:00:11.578704 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:f24234939afca841e46ea4d17bec959b63705ab0e75476465e777d44905c5f1b" Dec 04 18:00:11 crc kubenswrapper[4733]: E1204 18:00:11.578960 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:f24234939afca841e46ea4d17bec959b63705ab0e75476465e777d44905c5f1b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-krgsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-2khcb_openstack(8c53e91d-1256-408a-9b97-a048f425c441): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 18:00:11 crc kubenswrapper[4733]: E1204 18:00:11.580407 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-2khcb" podUID="8c53e91d-1256-408a-9b97-a048f425c441" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.663417 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.737401 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-swift-storage-0\") pod \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.737693 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-config\") pod \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.737874 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-nb\") pod \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.737954 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-svc\") pod \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.737984 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-sb\") pod \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.738013 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4ws6\" (UniqueName: \"kubernetes.io/projected/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-kube-api-access-j4ws6\") pod \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\" (UID: \"5fdc02f1-40d7-43a1-9869-7815e0e6d2db\") " Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.744432 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-kube-api-access-j4ws6" (OuterVolumeSpecName: "kube-api-access-j4ws6") pod "5fdc02f1-40d7-43a1-9869-7815e0e6d2db" (UID: "5fdc02f1-40d7-43a1-9869-7815e0e6d2db"). InnerVolumeSpecName "kube-api-access-j4ws6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.783918 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5fdc02f1-40d7-43a1-9869-7815e0e6d2db" (UID: "5fdc02f1-40d7-43a1-9869-7815e0e6d2db"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.784611 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5fdc02f1-40d7-43a1-9869-7815e0e6d2db" (UID: "5fdc02f1-40d7-43a1-9869-7815e0e6d2db"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.785354 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5fdc02f1-40d7-43a1-9869-7815e0e6d2db" (UID: "5fdc02f1-40d7-43a1-9869-7815e0e6d2db"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.789396 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5fdc02f1-40d7-43a1-9869-7815e0e6d2db" (UID: "5fdc02f1-40d7-43a1-9869-7815e0e6d2db"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.795129 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-config" (OuterVolumeSpecName: "config") pod "5fdc02f1-40d7-43a1-9869-7815e0e6d2db" (UID: "5fdc02f1-40d7-43a1-9869-7815e0e6d2db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.840538 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.840575 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.840586 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.840597 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4ws6\" (UniqueName: \"kubernetes.io/projected/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-kube-api-access-j4ws6\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.840608 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:11 crc kubenswrapper[4733]: I1204 18:00:11.840618 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fdc02f1-40d7-43a1-9869-7815e0e6d2db-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:12 crc kubenswrapper[4733]: E1204 18:00:12.152688 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f" Dec 04 18:00:12 crc kubenswrapper[4733]: E1204 18:00:12.153062 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9spkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-gbxkg_openstack(3be601db-3698-4272-8005-083d5b93cbfd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 18:00:12 crc kubenswrapper[4733]: E1204 18:00:12.154246 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-gbxkg" podUID="3be601db-3698-4272-8005-083d5b93cbfd" Dec 04 18:00:12 crc kubenswrapper[4733]: I1204 18:00:12.349243 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" Dec 04 18:00:12 crc kubenswrapper[4733]: I1204 18:00:12.349665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-779c5847bc-ptkcg" event={"ID":"5fdc02f1-40d7-43a1-9869-7815e0e6d2db","Type":"ContainerDied","Data":"a84adfb68199df745f8e243aeda49fb2c1210b88bba85d44083c058b1bb29308"} Dec 04 18:00:12 crc kubenswrapper[4733]: I1204 18:00:12.349711 4733 scope.go:117] "RemoveContainer" containerID="8d980a83191192ad31b4a99fa1e5f80399e038222d823f299b5b3eac1a730fda" Dec 04 18:00:12 crc kubenswrapper[4733]: E1204 18:00:12.350469 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:82006b9c64d4c5f80483cda262d960ce6be4813665158ef1a53ea7734bbe431f\\\"\"" pod="openstack/barbican-db-sync-gbxkg" podUID="3be601db-3698-4272-8005-083d5b93cbfd" Dec 04 18:00:12 crc kubenswrapper[4733]: E1204 18:00:12.352017 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:f24234939afca841e46ea4d17bec959b63705ab0e75476465e777d44905c5f1b\\\"\"" pod="openstack/placement-db-sync-2khcb" podUID="8c53e91d-1256-408a-9b97-a048f425c441" Dec 04 18:00:12 crc kubenswrapper[4733]: I1204 18:00:12.415548 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-ptkcg"] Dec 04 18:00:12 crc kubenswrapper[4733]: I1204 18:00:12.424946 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-779c5847bc-ptkcg"] Dec 04 18:00:14 crc kubenswrapper[4733]: I1204 18:00:14.352039 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" path="/var/lib/kubelet/pods/5fdc02f1-40d7-43a1-9869-7815e0e6d2db/volumes" Dec 04 18:00:14 crc kubenswrapper[4733]: I1204 18:00:14.551107 4733 scope.go:117] "RemoveContainer" containerID="89c5383d6b95f43e1f6d91b8ef881a2cf07c6e3ee81c158f06ced81158b76bd4" Dec 04 18:00:14 crc kubenswrapper[4733]: E1204 18:00:14.554457 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2" Dec 04 18:00:14 crc kubenswrapper[4733]: E1204 18:00:14.554680 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dkxzz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-btknr_openstack(3bcc7f96-8d09-4d23-8e53-02b679d7851a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 18:00:14 crc kubenswrapper[4733]: E1204 18:00:14.555950 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-btknr" podUID="3bcc7f96-8d09-4d23-8e53-02b679d7851a" Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.040911 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml"] Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.084524 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nkhjz"] Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.381447 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"abc26201-5065-4a19-9311-fc90d11401d0","Type":"ContainerStarted","Data":"d3243d7af0dfc453f4d69cd23ee6a59272dc78ba3c8ed8f8e372639c8284d8ae"} Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.385723 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkhjz" event={"ID":"2fca0e02-094d-4588-9efa-09181761155a","Type":"ContainerStarted","Data":"e9f80fba5d806616b88bd80d77acee2d8113ec26a9b590174f91a0f9d6ab8ba9"} Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.385754 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkhjz" event={"ID":"2fca0e02-094d-4588-9efa-09181761155a","Type":"ContainerStarted","Data":"e57040fa77ad83fa1149158fc5aa02312ed3c5a75fddc0ff475ce1655890b908"} Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.389341 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" event={"ID":"4ce93f96-1275-437f-9bcd-a7b299bae008","Type":"ContainerStarted","Data":"e5ddd5102295aabdfc76466c7e2b9419618ad91881c05eeb27ba6cac4a5c93ab"} Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.389396 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" event={"ID":"4ce93f96-1275-437f-9bcd-a7b299bae008","Type":"ContainerStarted","Data":"9371e5801cb54f78c04bad6a02954d9d9f71afb72cb5f5e8ff98309a83bb5347"} Dec 04 18:00:15 crc kubenswrapper[4733]: E1204 18:00:15.393178 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2\\\"\"" pod="openstack/cinder-db-sync-btknr" podUID="3bcc7f96-8d09-4d23-8e53-02b679d7851a" Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.410858 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nkhjz" podStartSLOduration=6.410842136 podStartE2EDuration="6.410842136s" podCreationTimestamp="2025-12-04 18:00:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:15.405595068 +0000 UTC m=+1277.360956114" watchObservedRunningTime="2025-12-04 18:00:15.410842136 +0000 UTC m=+1277.366203182" Dec 04 18:00:15 crc kubenswrapper[4733]: I1204 18:00:15.446460 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" podStartSLOduration=15.446443853 podStartE2EDuration="15.446443853s" podCreationTimestamp="2025-12-04 18:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:15.438928025 +0000 UTC m=+1277.394289071" watchObservedRunningTime="2025-12-04 18:00:15.446443853 +0000 UTC m=+1277.401804899" Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.412526 4733 generic.go:334] "Generic (PLEG): container finished" podID="4ce93f96-1275-437f-9bcd-a7b299bae008" containerID="e5ddd5102295aabdfc76466c7e2b9419618ad91881c05eeb27ba6cac4a5c93ab" exitCode=0 Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.412615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" event={"ID":"4ce93f96-1275-437f-9bcd-a7b299bae008","Type":"ContainerDied","Data":"e5ddd5102295aabdfc76466c7e2b9419618ad91881c05eeb27ba6cac4a5c93ab"} Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.421447 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"40314bda-46f4-426e-aeff-0029afdf5c90","Type":"ContainerStarted","Data":"643fab1fc6fa29e09ee7b9e662e869940d502d3a5ca9eea01011f6c047c63755"} Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.423020 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"40314bda-46f4-426e-aeff-0029afdf5c90","Type":"ContainerStarted","Data":"1471d44f86f056dac90145e65c91eb8e660b5f31b908dcb5f5b7cb79fc5a7064"} Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.421504 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" containerName="glance-log" containerID="cri-o://1471d44f86f056dac90145e65c91eb8e660b5f31b908dcb5f5b7cb79fc5a7064" gracePeriod=30 Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.421545 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" containerName="glance-httpd" containerID="cri-o://643fab1fc6fa29e09ee7b9e662e869940d502d3a5ca9eea01011f6c047c63755" gracePeriod=30 Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.432916 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51031e1a-06fe-441e-a9d7-97523825a26e","Type":"ContainerStarted","Data":"77118deac3a99d7641547823e8365e3ec7939353fa68ed8de8ec7575b41dd783"} Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.433119 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" containerName="glance-log" containerID="cri-o://8db2bb8eb14546ebea7175ee3d8b0c49509e28fb3250d5aa493e06d670d5943c" gracePeriod=30 Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.433160 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" containerName="glance-httpd" containerID="cri-o://77118deac3a99d7641547823e8365e3ec7939353fa68ed8de8ec7575b41dd783" gracePeriod=30 Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.489267 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=23.489246275 podStartE2EDuration="23.489246275s" podCreationTimestamp="2025-12-04 17:59:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:16.461080734 +0000 UTC m=+1278.416441790" watchObservedRunningTime="2025-12-04 18:00:16.489246275 +0000 UTC m=+1278.444607331" Dec 04 18:00:16 crc kubenswrapper[4733]: I1204 18:00:16.491641 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=23.491631978 podStartE2EDuration="23.491631978s" podCreationTimestamp="2025-12-04 17:59:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:16.485714432 +0000 UTC m=+1278.441075478" watchObservedRunningTime="2025-12-04 18:00:16.491631978 +0000 UTC m=+1278.446993024" Dec 04 18:00:17 crc kubenswrapper[4733]: I1204 18:00:17.448434 4733 generic.go:334] "Generic (PLEG): container finished" podID="40314bda-46f4-426e-aeff-0029afdf5c90" containerID="643fab1fc6fa29e09ee7b9e662e869940d502d3a5ca9eea01011f6c047c63755" exitCode=0 Dec 04 18:00:17 crc kubenswrapper[4733]: I1204 18:00:17.448724 4733 generic.go:334] "Generic (PLEG): container finished" podID="40314bda-46f4-426e-aeff-0029afdf5c90" containerID="1471d44f86f056dac90145e65c91eb8e660b5f31b908dcb5f5b7cb79fc5a7064" exitCode=143 Dec 04 18:00:17 crc kubenswrapper[4733]: I1204 18:00:17.448828 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"40314bda-46f4-426e-aeff-0029afdf5c90","Type":"ContainerDied","Data":"643fab1fc6fa29e09ee7b9e662e869940d502d3a5ca9eea01011f6c047c63755"} Dec 04 18:00:17 crc kubenswrapper[4733]: I1204 18:00:17.448878 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"40314bda-46f4-426e-aeff-0029afdf5c90","Type":"ContainerDied","Data":"1471d44f86f056dac90145e65c91eb8e660b5f31b908dcb5f5b7cb79fc5a7064"} Dec 04 18:00:17 crc kubenswrapper[4733]: I1204 18:00:17.451713 4733 generic.go:334] "Generic (PLEG): container finished" podID="51031e1a-06fe-441e-a9d7-97523825a26e" containerID="77118deac3a99d7641547823e8365e3ec7939353fa68ed8de8ec7575b41dd783" exitCode=0 Dec 04 18:00:17 crc kubenswrapper[4733]: I1204 18:00:17.451745 4733 generic.go:334] "Generic (PLEG): container finished" podID="51031e1a-06fe-441e-a9d7-97523825a26e" containerID="8db2bb8eb14546ebea7175ee3d8b0c49509e28fb3250d5aa493e06d670d5943c" exitCode=143 Dec 04 18:00:17 crc kubenswrapper[4733]: I1204 18:00:17.451783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51031e1a-06fe-441e-a9d7-97523825a26e","Type":"ContainerDied","Data":"77118deac3a99d7641547823e8365e3ec7939353fa68ed8de8ec7575b41dd783"} Dec 04 18:00:17 crc kubenswrapper[4733]: I1204 18:00:17.451837 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51031e1a-06fe-441e-a9d7-97523825a26e","Type":"ContainerDied","Data":"8db2bb8eb14546ebea7175ee3d8b0c49509e28fb3250d5aa493e06d670d5943c"} Dec 04 18:00:18 crc kubenswrapper[4733]: I1204 18:00:18.461041 4733 generic.go:334] "Generic (PLEG): container finished" podID="2fca0e02-094d-4588-9efa-09181761155a" containerID="e9f80fba5d806616b88bd80d77acee2d8113ec26a9b590174f91a0f9d6ab8ba9" exitCode=0 Dec 04 18:00:18 crc kubenswrapper[4733]: I1204 18:00:18.461219 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkhjz" event={"ID":"2fca0e02-094d-4588-9efa-09181761155a","Type":"ContainerDied","Data":"e9f80fba5d806616b88bd80d77acee2d8113ec26a9b590174f91a0f9d6ab8ba9"} Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.487466 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"40314bda-46f4-426e-aeff-0029afdf5c90","Type":"ContainerDied","Data":"7e20574e0ea3bc22f3ad1eff735250e2f9c5e96909cab5495e5a236a09860624"} Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.487974 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e20574e0ea3bc22f3ad1eff735250e2f9c5e96909cab5495e5a236a09860624" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.493282 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51031e1a-06fe-441e-a9d7-97523825a26e","Type":"ContainerDied","Data":"7a6dff60f8934ffdbe65b066586d3087c7fdc86a69c97b91582b969f4bd58efc"} Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.493406 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a6dff60f8934ffdbe65b066586d3087c7fdc86a69c97b91582b969f4bd58efc" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.495578 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" event={"ID":"4ce93f96-1275-437f-9bcd-a7b299bae008","Type":"ContainerDied","Data":"9371e5801cb54f78c04bad6a02954d9d9f71afb72cb5f5e8ff98309a83bb5347"} Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.495607 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9371e5801cb54f78c04bad6a02954d9d9f71afb72cb5f5e8ff98309a83bb5347" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.498775 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nkhjz" event={"ID":"2fca0e02-094d-4588-9efa-09181761155a","Type":"ContainerDied","Data":"e57040fa77ad83fa1149158fc5aa02312ed3c5a75fddc0ff475ce1655890b908"} Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.498929 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e57040fa77ad83fa1149158fc5aa02312ed3c5a75fddc0ff475ce1655890b908" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.641257 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.660396 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.719895 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738300 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-scripts\") pod \"2fca0e02-094d-4588-9efa-09181761155a\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738382 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-577s5\" (UniqueName: \"kubernetes.io/projected/2fca0e02-094d-4588-9efa-09181761155a-kube-api-access-577s5\") pod \"2fca0e02-094d-4588-9efa-09181761155a\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738455 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-combined-ca-bundle\") pod \"2fca0e02-094d-4588-9efa-09181761155a\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738490 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"51031e1a-06fe-441e-a9d7-97523825a26e\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738587 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vm97\" (UniqueName: \"kubernetes.io/projected/51031e1a-06fe-441e-a9d7-97523825a26e-kube-api-access-5vm97\") pod \"51031e1a-06fe-441e-a9d7-97523825a26e\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738698 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-credential-keys\") pod \"2fca0e02-094d-4588-9efa-09181761155a\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738743 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-fernet-keys\") pod \"2fca0e02-094d-4588-9efa-09181761155a\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738790 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zczw9\" (UniqueName: \"kubernetes.io/projected/4ce93f96-1275-437f-9bcd-a7b299bae008-kube-api-access-zczw9\") pod \"4ce93f96-1275-437f-9bcd-a7b299bae008\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738849 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-scripts\") pod \"51031e1a-06fe-441e-a9d7-97523825a26e\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738886 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-config-data\") pod \"2fca0e02-094d-4588-9efa-09181761155a\" (UID: \"2fca0e02-094d-4588-9efa-09181761155a\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.738985 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-combined-ca-bundle\") pod \"51031e1a-06fe-441e-a9d7-97523825a26e\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.739017 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-httpd-run\") pod \"51031e1a-06fe-441e-a9d7-97523825a26e\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.739071 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce93f96-1275-437f-9bcd-a7b299bae008-config-volume\") pod \"4ce93f96-1275-437f-9bcd-a7b299bae008\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.739131 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-logs\") pod \"51031e1a-06fe-441e-a9d7-97523825a26e\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.739159 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-config-data\") pod \"51031e1a-06fe-441e-a9d7-97523825a26e\" (UID: \"51031e1a-06fe-441e-a9d7-97523825a26e\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.739260 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce93f96-1275-437f-9bcd-a7b299bae008-secret-volume\") pod \"4ce93f96-1275-437f-9bcd-a7b299bae008\" (UID: \"4ce93f96-1275-437f-9bcd-a7b299bae008\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.747056 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "51031e1a-06fe-441e-a9d7-97523825a26e" (UID: "51031e1a-06fe-441e-a9d7-97523825a26e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.757827 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ce93f96-1275-437f-9bcd-a7b299bae008-kube-api-access-zczw9" (OuterVolumeSpecName: "kube-api-access-zczw9") pod "4ce93f96-1275-437f-9bcd-a7b299bae008" (UID: "4ce93f96-1275-437f-9bcd-a7b299bae008"). InnerVolumeSpecName "kube-api-access-zczw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.758023 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce93f96-1275-437f-9bcd-a7b299bae008-config-volume" (OuterVolumeSpecName: "config-volume") pod "4ce93f96-1275-437f-9bcd-a7b299bae008" (UID: "4ce93f96-1275-437f-9bcd-a7b299bae008"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.758404 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-scripts" (OuterVolumeSpecName: "scripts") pod "2fca0e02-094d-4588-9efa-09181761155a" (UID: "2fca0e02-094d-4588-9efa-09181761155a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.758740 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.758902 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce93f96-1275-437f-9bcd-a7b299bae008-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4ce93f96-1275-437f-9bcd-a7b299bae008" (UID: "4ce93f96-1275-437f-9bcd-a7b299bae008"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.762829 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-logs" (OuterVolumeSpecName: "logs") pod "51031e1a-06fe-441e-a9d7-97523825a26e" (UID: "51031e1a-06fe-441e-a9d7-97523825a26e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.763394 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2fca0e02-094d-4588-9efa-09181761155a" (UID: "2fca0e02-094d-4588-9efa-09181761155a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.765031 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51031e1a-06fe-441e-a9d7-97523825a26e-kube-api-access-5vm97" (OuterVolumeSpecName: "kube-api-access-5vm97") pod "51031e1a-06fe-441e-a9d7-97523825a26e" (UID: "51031e1a-06fe-441e-a9d7-97523825a26e"). InnerVolumeSpecName "kube-api-access-5vm97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.766682 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2fca0e02-094d-4588-9efa-09181761155a" (UID: "2fca0e02-094d-4588-9efa-09181761155a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.768999 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-scripts" (OuterVolumeSpecName: "scripts") pod "51031e1a-06fe-441e-a9d7-97523825a26e" (UID: "51031e1a-06fe-441e-a9d7-97523825a26e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.789025 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fca0e02-094d-4588-9efa-09181761155a-kube-api-access-577s5" (OuterVolumeSpecName: "kube-api-access-577s5") pod "2fca0e02-094d-4588-9efa-09181761155a" (UID: "2fca0e02-094d-4588-9efa-09181761155a"). InnerVolumeSpecName "kube-api-access-577s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.790952 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "51031e1a-06fe-441e-a9d7-97523825a26e" (UID: "51031e1a-06fe-441e-a9d7-97523825a26e"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.813637 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fca0e02-094d-4588-9efa-09181761155a" (UID: "2fca0e02-094d-4588-9efa-09181761155a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.840461 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-config-data\") pod \"40314bda-46f4-426e-aeff-0029afdf5c90\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.840512 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-logs\") pod \"40314bda-46f4-426e-aeff-0029afdf5c90\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.840556 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-scripts\") pod \"40314bda-46f4-426e-aeff-0029afdf5c90\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.840594 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-httpd-run\") pod \"40314bda-46f4-426e-aeff-0029afdf5c90\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.840676 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-combined-ca-bundle\") pod \"40314bda-46f4-426e-aeff-0029afdf5c90\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.840742 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"40314bda-46f4-426e-aeff-0029afdf5c90\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.840777 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc44h\" (UniqueName: \"kubernetes.io/projected/40314bda-46f4-426e-aeff-0029afdf5c90-kube-api-access-zc44h\") pod \"40314bda-46f4-426e-aeff-0029afdf5c90\" (UID: \"40314bda-46f4-426e-aeff-0029afdf5c90\") " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841273 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841289 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce93f96-1275-437f-9bcd-a7b299bae008-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841303 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51031e1a-06fe-441e-a9d7-97523825a26e-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841313 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce93f96-1275-437f-9bcd-a7b299bae008-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841324 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841335 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-577s5\" (UniqueName: \"kubernetes.io/projected/2fca0e02-094d-4588-9efa-09181761155a-kube-api-access-577s5\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841346 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841369 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841380 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vm97\" (UniqueName: \"kubernetes.io/projected/51031e1a-06fe-441e-a9d7-97523825a26e-kube-api-access-5vm97\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841391 4733 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841439 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841451 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zczw9\" (UniqueName: \"kubernetes.io/projected/4ce93f96-1275-437f-9bcd-a7b299bae008-kube-api-access-zczw9\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.841462 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.843487 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-logs" (OuterVolumeSpecName: "logs") pod "40314bda-46f4-426e-aeff-0029afdf5c90" (UID: "40314bda-46f4-426e-aeff-0029afdf5c90"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.843512 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "40314bda-46f4-426e-aeff-0029afdf5c90" (UID: "40314bda-46f4-426e-aeff-0029afdf5c90"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.850607 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-scripts" (OuterVolumeSpecName: "scripts") pod "40314bda-46f4-426e-aeff-0029afdf5c90" (UID: "40314bda-46f4-426e-aeff-0029afdf5c90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.855920 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40314bda-46f4-426e-aeff-0029afdf5c90-kube-api-access-zc44h" (OuterVolumeSpecName: "kube-api-access-zc44h") pod "40314bda-46f4-426e-aeff-0029afdf5c90" (UID: "40314bda-46f4-426e-aeff-0029afdf5c90"). InnerVolumeSpecName "kube-api-access-zc44h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.856129 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "40314bda-46f4-426e-aeff-0029afdf5c90" (UID: "40314bda-46f4-426e-aeff-0029afdf5c90"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.872977 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.878166 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40314bda-46f4-426e-aeff-0029afdf5c90" (UID: "40314bda-46f4-426e-aeff-0029afdf5c90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.881562 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-config-data" (OuterVolumeSpecName: "config-data") pod "2fca0e02-094d-4588-9efa-09181761155a" (UID: "2fca0e02-094d-4588-9efa-09181761155a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.884472 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51031e1a-06fe-441e-a9d7-97523825a26e" (UID: "51031e1a-06fe-441e-a9d7-97523825a26e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.904985 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-config-data" (OuterVolumeSpecName: "config-data") pod "40314bda-46f4-426e-aeff-0029afdf5c90" (UID: "40314bda-46f4-426e-aeff-0029afdf5c90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.909847 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-config-data" (OuterVolumeSpecName: "config-data") pod "51031e1a-06fe-441e-a9d7-97523825a26e" (UID: "51031e1a-06fe-441e-a9d7-97523825a26e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943513 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc44h\" (UniqueName: \"kubernetes.io/projected/40314bda-46f4-426e-aeff-0029afdf5c90-kube-api-access-zc44h\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943607 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fca0e02-094d-4588-9efa-09181761155a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943623 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943633 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943653 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943661 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943669 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/40314bda-46f4-426e-aeff-0029afdf5c90-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943677 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51031e1a-06fe-441e-a9d7-97523825a26e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943686 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40314bda-46f4-426e-aeff-0029afdf5c90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943696 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.943731 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 04 18:00:21 crc kubenswrapper[4733]: I1204 18:00:21.967303 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.045185 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.512589 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nkhjz" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.512588 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"abc26201-5065-4a19-9311-fc90d11401d0","Type":"ContainerStarted","Data":"4a19da6f8268813a440ce518829aa523a1c4aa0a03e0e08d93c751d82e5bf4e0"} Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.512631 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.512696 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.512721 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.555174 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.563309 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.575842 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.582047 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.589982 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:00:22 crc kubenswrapper[4733]: E1204 18:00:22.590303 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" containerName="glance-log" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590316 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" containerName="glance-log" Dec 04 18:00:22 crc kubenswrapper[4733]: E1204 18:00:22.590329 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce93f96-1275-437f-9bcd-a7b299bae008" containerName="collect-profiles" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590335 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce93f96-1275-437f-9bcd-a7b299bae008" containerName="collect-profiles" Dec 04 18:00:22 crc kubenswrapper[4733]: E1204 18:00:22.590349 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" containerName="glance-httpd" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590354 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" containerName="glance-httpd" Dec 04 18:00:22 crc kubenswrapper[4733]: E1204 18:00:22.590367 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" containerName="glance-log" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590375 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" containerName="glance-log" Dec 04 18:00:22 crc kubenswrapper[4733]: E1204 18:00:22.590388 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" containerName="glance-httpd" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590394 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" containerName="glance-httpd" Dec 04 18:00:22 crc kubenswrapper[4733]: E1204 18:00:22.590404 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fca0e02-094d-4588-9efa-09181761155a" containerName="keystone-bootstrap" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590410 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fca0e02-094d-4588-9efa-09181761155a" containerName="keystone-bootstrap" Dec 04 18:00:22 crc kubenswrapper[4733]: E1204 18:00:22.590487 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerName="dnsmasq-dns" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590494 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerName="dnsmasq-dns" Dec 04 18:00:22 crc kubenswrapper[4733]: E1204 18:00:22.590512 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerName="init" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590518 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerName="init" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590670 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" containerName="glance-httpd" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590681 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" containerName="glance-log" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590695 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" containerName="glance-httpd" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590706 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" containerName="glance-log" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590713 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fca0e02-094d-4588-9efa-09181761155a" containerName="keystone-bootstrap" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590723 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fdc02f1-40d7-43a1-9869-7815e0e6d2db" containerName="dnsmasq-dns" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.590731 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce93f96-1275-437f-9bcd-a7b299bae008" containerName="collect-profiles" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.591504 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.595386 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.595736 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.596003 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-g46cp" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.596203 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.600357 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.610186 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.611553 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.613096 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.613560 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.643936 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.656008 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.656122 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.656454 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.656529 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.656560 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.656577 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.656600 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.656636 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpfbg\" (UniqueName: \"kubernetes.io/projected/f4f82a6d-dcad-432e-a365-38df92ff4e6e-kube-api-access-dpfbg\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758517 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf8hb\" (UniqueName: \"kubernetes.io/projected/527118fc-11ba-4c6b-a15a-415912a5e650-kube-api-access-cf8hb\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758574 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-scripts\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758649 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758678 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758745 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758809 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758847 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758909 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-config-data\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758941 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.758977 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759025 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759068 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-logs\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759092 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759121 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpfbg\" (UniqueName: \"kubernetes.io/projected/f4f82a6d-dcad-432e-a365-38df92ff4e6e-kube-api-access-dpfbg\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759167 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759294 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759344 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759410 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.759736 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.763171 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.763429 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.763949 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.764329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.778746 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpfbg\" (UniqueName: \"kubernetes.io/projected/f4f82a6d-dcad-432e-a365-38df92ff4e6e-kube-api-access-dpfbg\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.791352 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860071 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf8hb\" (UniqueName: \"kubernetes.io/projected/527118fc-11ba-4c6b-a15a-415912a5e650-kube-api-access-cf8hb\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860112 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-scripts\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860135 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860179 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860209 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-config-data\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860241 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-logs\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860270 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860298 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860754 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.860780 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.861058 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-logs\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.865065 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.868517 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-scripts\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.871254 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-config-data\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.872263 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.882946 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf8hb\" (UniqueName: \"kubernetes.io/projected/527118fc-11ba-4c6b-a15a-415912a5e650-kube-api-access-cf8hb\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.896987 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.917111 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.933149 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-64bd4cdd64-fftgx"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.934741 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.935252 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.938930 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.939121 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.939771 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.944849 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64bd4cdd64-fftgx"] Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.958314 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.958809 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r9kr4" Dec 04 18:00:22 crc kubenswrapper[4733]: I1204 18:00:22.959076 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.073523 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-internal-tls-certs\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.073865 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-config-data\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.073891 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-public-tls-certs\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.073921 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-credential-keys\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.073946 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-combined-ca-bundle\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.074008 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-fernet-keys\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.074035 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-scripts\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.074123 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2jgd\" (UniqueName: \"kubernetes.io/projected/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-kube-api-access-w2jgd\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.175104 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-internal-tls-certs\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.175152 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-config-data\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.175176 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-public-tls-certs\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.175196 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-credential-keys\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.175872 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-combined-ca-bundle\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.175947 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-fernet-keys\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.176280 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-scripts\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.176357 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2jgd\" (UniqueName: \"kubernetes.io/projected/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-kube-api-access-w2jgd\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.181292 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-public-tls-certs\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.181306 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-fernet-keys\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.181608 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-credential-keys\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.182366 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-combined-ca-bundle\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.184847 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-scripts\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.186571 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-config-data\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.188361 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-internal-tls-certs\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.200087 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2jgd\" (UniqueName: \"kubernetes.io/projected/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-kube-api-access-w2jgd\") pod \"keystone-64bd4cdd64-fftgx\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.295419 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.546182 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:00:23 crc kubenswrapper[4733]: W1204 18:00:23.564609 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f82a6d_dcad_432e_a365_38df92ff4e6e.slice/crio-84156c3fe19ccf80726b8ee78e8b0c7f64e95fe849240198a520dced13ef8220 WatchSource:0}: Error finding container 84156c3fe19ccf80726b8ee78e8b0c7f64e95fe849240198a520dced13ef8220: Status 404 returned error can't find the container with id 84156c3fe19ccf80726b8ee78e8b0c7f64e95fe849240198a520dced13ef8220 Dec 04 18:00:23 crc kubenswrapper[4733]: I1204 18:00:23.755513 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64bd4cdd64-fftgx"] Dec 04 18:00:23 crc kubenswrapper[4733]: W1204 18:00:23.776066 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9dd2ad1b_c8c0_4bdb_b46c_4bfd8d3fe1fa.slice/crio-d401f46e77cc3710d349a561b1377b9d2083e6630dd963949302bf71530d8dfb WatchSource:0}: Error finding container d401f46e77cc3710d349a561b1377b9d2083e6630dd963949302bf71530d8dfb: Status 404 returned error can't find the container with id d401f46e77cc3710d349a561b1377b9d2083e6630dd963949302bf71530d8dfb Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.345254 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40314bda-46f4-426e-aeff-0029afdf5c90" path="/var/lib/kubelet/pods/40314bda-46f4-426e-aeff-0029afdf5c90/volumes" Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.346462 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51031e1a-06fe-441e-a9d7-97523825a26e" path="/var/lib/kubelet/pods/51031e1a-06fe-441e-a9d7-97523825a26e/volumes" Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.541417 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4f82a6d-dcad-432e-a365-38df92ff4e6e","Type":"ContainerStarted","Data":"1241cadb321985565a1841b5836435da7c72c4337f283fd97ea244d6cfe421d2"} Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.542168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4f82a6d-dcad-432e-a365-38df92ff4e6e","Type":"ContainerStarted","Data":"84156c3fe19ccf80726b8ee78e8b0c7f64e95fe849240198a520dced13ef8220"} Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.543052 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64bd4cdd64-fftgx" event={"ID":"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa","Type":"ContainerStarted","Data":"912469a5b6cd827fb51c283e6484beba8f59e32cbc52ed716d042adc91ed7c90"} Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.543081 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64bd4cdd64-fftgx" event={"ID":"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa","Type":"ContainerStarted","Data":"d401f46e77cc3710d349a561b1377b9d2083e6630dd963949302bf71530d8dfb"} Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.543215 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.558977 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-64bd4cdd64-fftgx" podStartSLOduration=2.558958692 podStartE2EDuration="2.558958692s" podCreationTimestamp="2025-12-04 18:00:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:24.557299228 +0000 UTC m=+1286.512660294" watchObservedRunningTime="2025-12-04 18:00:24.558958692 +0000 UTC m=+1286.514319738" Dec 04 18:00:24 crc kubenswrapper[4733]: I1204 18:00:24.626970 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:00:24 crc kubenswrapper[4733]: W1204 18:00:24.630277 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod527118fc_11ba_4c6b_a15a_415912a5e650.slice/crio-5ef2507be5bdc6228b6a16c06d1aaf7a8d712c4efea9ff32778800be6bbef7c3 WatchSource:0}: Error finding container 5ef2507be5bdc6228b6a16c06d1aaf7a8d712c4efea9ff32778800be6bbef7c3: Status 404 returned error can't find the container with id 5ef2507be5bdc6228b6a16c06d1aaf7a8d712c4efea9ff32778800be6bbef7c3 Dec 04 18:00:25 crc kubenswrapper[4733]: I1204 18:00:25.550982 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"527118fc-11ba-4c6b-a15a-415912a5e650","Type":"ContainerStarted","Data":"5ef2507be5bdc6228b6a16c06d1aaf7a8d712c4efea9ff32778800be6bbef7c3"} Dec 04 18:00:26 crc kubenswrapper[4733]: I1204 18:00:26.569361 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4f82a6d-dcad-432e-a365-38df92ff4e6e","Type":"ContainerStarted","Data":"5c0ddd09e462eda1e481a8b62bdcbb3fee36cd52d17f119c18c2049111ec6377"} Dec 04 18:00:26 crc kubenswrapper[4733]: I1204 18:00:26.571895 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"527118fc-11ba-4c6b-a15a-415912a5e650","Type":"ContainerStarted","Data":"db6f3a34d675cc354963a482ff62b85e52c39c97b79d0c7f54bfdaf0383617c0"} Dec 04 18:00:26 crc kubenswrapper[4733]: I1204 18:00:26.596624 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.5966047450000005 podStartE2EDuration="4.596604745s" podCreationTimestamp="2025-12-04 18:00:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:26.587490176 +0000 UTC m=+1288.542851222" watchObservedRunningTime="2025-12-04 18:00:26.596604745 +0000 UTC m=+1288.551965791" Dec 04 18:00:27 crc kubenswrapper[4733]: I1204 18:00:27.590186 4733 generic.go:334] "Generic (PLEG): container finished" podID="77bd0d5e-6bf1-4257-b30e-795688595c1c" containerID="d49942198858476882af324869180b206cbbcac041cbb1d3b56286f2de3f2d02" exitCode=0 Dec 04 18:00:27 crc kubenswrapper[4733]: I1204 18:00:27.590286 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2thjt" event={"ID":"77bd0d5e-6bf1-4257-b30e-795688595c1c","Type":"ContainerDied","Data":"d49942198858476882af324869180b206cbbcac041cbb1d3b56286f2de3f2d02"} Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.075241 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2thjt" Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.206075 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzx8t\" (UniqueName: \"kubernetes.io/projected/77bd0d5e-6bf1-4257-b30e-795688595c1c-kube-api-access-tzx8t\") pod \"77bd0d5e-6bf1-4257-b30e-795688595c1c\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.206334 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-config\") pod \"77bd0d5e-6bf1-4257-b30e-795688595c1c\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.206447 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-combined-ca-bundle\") pod \"77bd0d5e-6bf1-4257-b30e-795688595c1c\" (UID: \"77bd0d5e-6bf1-4257-b30e-795688595c1c\") " Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.217264 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77bd0d5e-6bf1-4257-b30e-795688595c1c-kube-api-access-tzx8t" (OuterVolumeSpecName: "kube-api-access-tzx8t") pod "77bd0d5e-6bf1-4257-b30e-795688595c1c" (UID: "77bd0d5e-6bf1-4257-b30e-795688595c1c"). InnerVolumeSpecName "kube-api-access-tzx8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.232830 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-config" (OuterVolumeSpecName: "config") pod "77bd0d5e-6bf1-4257-b30e-795688595c1c" (UID: "77bd0d5e-6bf1-4257-b30e-795688595c1c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.243302 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77bd0d5e-6bf1-4257-b30e-795688595c1c" (UID: "77bd0d5e-6bf1-4257-b30e-795688595c1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.308818 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.308850 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bd0d5e-6bf1-4257-b30e-795688595c1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.308861 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzx8t\" (UniqueName: \"kubernetes.io/projected/77bd0d5e-6bf1-4257-b30e-795688595c1c-kube-api-access-tzx8t\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.622228 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2thjt" event={"ID":"77bd0d5e-6bf1-4257-b30e-795688595c1c","Type":"ContainerDied","Data":"670ab638276cc50e23ac089ddd4b8e12b115ef64fd8fa0dbef395a8460dc8f39"} Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.622273 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="670ab638276cc50e23ac089ddd4b8e12b115ef64fd8fa0dbef395a8460dc8f39" Dec 04 18:00:30 crc kubenswrapper[4733]: I1204 18:00:30.622342 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2thjt" Dec 04 18:00:30 crc kubenswrapper[4733]: E1204 18:00:30.806634 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="abc26201-5065-4a19-9311-fc90d11401d0" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.217370 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-rms44"] Dec 04 18:00:31 crc kubenswrapper[4733]: E1204 18:00:31.218026 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77bd0d5e-6bf1-4257-b30e-795688595c1c" containerName="neutron-db-sync" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.218039 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="77bd0d5e-6bf1-4257-b30e-795688595c1c" containerName="neutron-db-sync" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.218225 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="77bd0d5e-6bf1-4257-b30e-795688595c1c" containerName="neutron-db-sync" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.224138 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.249485 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-rms44"] Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.327963 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz8jz\" (UniqueName: \"kubernetes.io/projected/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-kube-api-access-bz8jz\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.328017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-config\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.328079 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-svc\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.328100 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.328142 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.328168 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.429996 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz8jz\" (UniqueName: \"kubernetes.io/projected/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-kube-api-access-bz8jz\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.430068 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-config\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.430149 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-svc\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.430186 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.430249 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.430286 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.431449 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-sb\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.433959 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-config\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.434093 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-svc\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.434331 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-swift-storage-0\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.434506 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-nb\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.456066 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz8jz\" (UniqueName: \"kubernetes.io/projected/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-kube-api-access-bz8jz\") pod \"dnsmasq-dns-77f55878d5-rms44\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.477647 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-98f7dbcd8-vdtg8"] Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.479357 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.482214 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.482422 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rhhsz" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.482626 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.482783 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.502498 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-98f7dbcd8-vdtg8"] Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.587299 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.632847 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxzdf\" (UniqueName: \"kubernetes.io/projected/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-kube-api-access-kxzdf\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.633269 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-ovndb-tls-certs\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.633349 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-config\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.633421 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-httpd-config\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.633478 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-combined-ca-bundle\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.637591 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"527118fc-11ba-4c6b-a15a-415912a5e650","Type":"ContainerStarted","Data":"ec9864790923d6c96b26352e7b30f7d1edab920785e6bd453aae8436fb4d67ab"} Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.644432 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"abc26201-5065-4a19-9311-fc90d11401d0","Type":"ContainerStarted","Data":"3894a7ff062a68f4a396cc9071b940d12d7b5f8e9e7fc8deb15155cc479476d6"} Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.644615 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="ceilometer-notification-agent" containerID="cri-o://d3243d7af0dfc453f4d69cd23ee6a59272dc78ba3c8ed8f8e372639c8284d8ae" gracePeriod=30 Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.644868 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.644920 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="proxy-httpd" containerID="cri-o://3894a7ff062a68f4a396cc9071b940d12d7b5f8e9e7fc8deb15155cc479476d6" gracePeriod=30 Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.644974 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="sg-core" containerID="cri-o://4a19da6f8268813a440ce518829aa523a1c4aa0a03e0e08d93c751d82e5bf4e0" gracePeriod=30 Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.659474 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2khcb" event={"ID":"8c53e91d-1256-408a-9b97-a048f425c441","Type":"ContainerStarted","Data":"9d57e43da79ab2c5d0b9a9159e21dc1548b4a16f3e0e5f81bcbb1076708a59e7"} Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.666949 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.666925638 podStartE2EDuration="9.666925638s" podCreationTimestamp="2025-12-04 18:00:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:31.659863672 +0000 UTC m=+1293.615224728" watchObservedRunningTime="2025-12-04 18:00:31.666925638 +0000 UTC m=+1293.622286684" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.673293 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-btknr" event={"ID":"3bcc7f96-8d09-4d23-8e53-02b679d7851a","Type":"ContainerStarted","Data":"03be397f987be20236ee4d9b4181dbb81db3d0b00ead6b8a03cfb6abed63df74"} Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.677733 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gbxkg" event={"ID":"3be601db-3698-4272-8005-083d5b93cbfd","Type":"ContainerStarted","Data":"e672d46632a950c5a99212c64ced6adecc83268c0542e4914c98073746bc1f0d"} Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.730737 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-2khcb" podStartSLOduration=2.910796563 podStartE2EDuration="42.730717566s" podCreationTimestamp="2025-12-04 17:59:49 +0000 UTC" firstStartedPulling="2025-12-04 17:59:50.789294281 +0000 UTC m=+1252.744655327" lastFinishedPulling="2025-12-04 18:00:30.609215274 +0000 UTC m=+1292.564576330" observedRunningTime="2025-12-04 18:00:31.727633725 +0000 UTC m=+1293.682994781" watchObservedRunningTime="2025-12-04 18:00:31.730717566 +0000 UTC m=+1293.686078622" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.734981 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-ovndb-tls-certs\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.735034 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-config\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.735055 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-httpd-config\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.735076 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-combined-ca-bundle\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.735100 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxzdf\" (UniqueName: \"kubernetes.io/projected/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-kube-api-access-kxzdf\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.745627 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-ovndb-tls-certs\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.746332 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-combined-ca-bundle\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.760857 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-httpd-config\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.761828 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-config\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.770187 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-gbxkg" podStartSLOduration=3.166435466 podStartE2EDuration="42.770161443s" podCreationTimestamp="2025-12-04 17:59:49 +0000 UTC" firstStartedPulling="2025-12-04 17:59:51.002893958 +0000 UTC m=+1252.958255004" lastFinishedPulling="2025-12-04 18:00:30.606619935 +0000 UTC m=+1292.561980981" observedRunningTime="2025-12-04 18:00:31.761137525 +0000 UTC m=+1293.716498571" watchObservedRunningTime="2025-12-04 18:00:31.770161443 +0000 UTC m=+1293.725522489" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.798553 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-btknr" podStartSLOduration=2.771946412 podStartE2EDuration="42.798530449s" podCreationTimestamp="2025-12-04 17:59:49 +0000 UTC" firstStartedPulling="2025-12-04 17:59:50.579198186 +0000 UTC m=+1252.534559232" lastFinishedPulling="2025-12-04 18:00:30.605782213 +0000 UTC m=+1292.561143269" observedRunningTime="2025-12-04 18:00:31.781911452 +0000 UTC m=+1293.737272508" watchObservedRunningTime="2025-12-04 18:00:31.798530449 +0000 UTC m=+1293.753891495" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.801107 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxzdf\" (UniqueName: \"kubernetes.io/projected/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-kube-api-access-kxzdf\") pod \"neutron-98f7dbcd8-vdtg8\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:31 crc kubenswrapper[4733]: I1204 18:00:31.825082 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.121673 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-rms44"] Dec 04 18:00:32 crc kubenswrapper[4733]: W1204 18:00:32.157524 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9c17e0c_1bb9_4532_9af6_4e46aac1227e.slice/crio-07b2126c667bcfcd15c4bef866e469414ba38eeee30cc6b242cd9cfe4cb68983 WatchSource:0}: Error finding container 07b2126c667bcfcd15c4bef866e469414ba38eeee30cc6b242cd9cfe4cb68983: Status 404 returned error can't find the container with id 07b2126c667bcfcd15c4bef866e469414ba38eeee30cc6b242cd9cfe4cb68983 Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.501964 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-98f7dbcd8-vdtg8"] Dec 04 18:00:32 crc kubenswrapper[4733]: W1204 18:00:32.505658 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bc84a76_c1da_4cb4_ba16_a3aeac81c1ab.slice/crio-a2e6fac54a9ac5724cc35da16d710ef5e2a58c20e7b09bd709b6a6a77ce42cef WatchSource:0}: Error finding container a2e6fac54a9ac5724cc35da16d710ef5e2a58c20e7b09bd709b6a6a77ce42cef: Status 404 returned error can't find the container with id a2e6fac54a9ac5724cc35da16d710ef5e2a58c20e7b09bd709b6a6a77ce42cef Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.687895 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-98f7dbcd8-vdtg8" event={"ID":"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab","Type":"ContainerStarted","Data":"a2e6fac54a9ac5724cc35da16d710ef5e2a58c20e7b09bd709b6a6a77ce42cef"} Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.689951 4733 generic.go:334] "Generic (PLEG): container finished" podID="abc26201-5065-4a19-9311-fc90d11401d0" containerID="3894a7ff062a68f4a396cc9071b940d12d7b5f8e9e7fc8deb15155cc479476d6" exitCode=0 Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.689989 4733 generic.go:334] "Generic (PLEG): container finished" podID="abc26201-5065-4a19-9311-fc90d11401d0" containerID="4a19da6f8268813a440ce518829aa523a1c4aa0a03e0e08d93c751d82e5bf4e0" exitCode=2 Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.690035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"abc26201-5065-4a19-9311-fc90d11401d0","Type":"ContainerDied","Data":"3894a7ff062a68f4a396cc9071b940d12d7b5f8e9e7fc8deb15155cc479476d6"} Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.690073 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"abc26201-5065-4a19-9311-fc90d11401d0","Type":"ContainerDied","Data":"4a19da6f8268813a440ce518829aa523a1c4aa0a03e0e08d93c751d82e5bf4e0"} Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.691710 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-rms44" event={"ID":"a9c17e0c-1bb9-4532-9af6-4e46aac1227e","Type":"ContainerStarted","Data":"3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b"} Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.691745 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-rms44" event={"ID":"a9c17e0c-1bb9-4532-9af6-4e46aac1227e","Type":"ContainerStarted","Data":"07b2126c667bcfcd15c4bef866e469414ba38eeee30cc6b242cd9cfe4cb68983"} Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.918847 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.919209 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.938966 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.939003 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.963329 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.981589 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.981750 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:32 crc kubenswrapper[4733]: I1204 18:00:32.992035 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.708616 4733 generic.go:334] "Generic (PLEG): container finished" podID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerID="3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b" exitCode=0 Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.708686 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-rms44" event={"ID":"a9c17e0c-1bb9-4532-9af6-4e46aac1227e","Type":"ContainerDied","Data":"3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b"} Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.717878 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-98f7dbcd8-vdtg8" event={"ID":"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab","Type":"ContainerStarted","Data":"5ae8831a27bc1dd563aa1309ef400056f32ce6d82dc70a522d2f89456bde35d9"} Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.717942 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-98f7dbcd8-vdtg8" event={"ID":"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab","Type":"ContainerStarted","Data":"262ab0c4956d796d1019c1aab9759043a7cd8f1e971e1c279de42fb81935131d"} Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.718021 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.718458 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.718569 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.718770 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.718789 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 18:00:33 crc kubenswrapper[4733]: I1204 18:00:33.781936 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-98f7dbcd8-vdtg8" podStartSLOduration=2.781918386 podStartE2EDuration="2.781918386s" podCreationTimestamp="2025-12-04 18:00:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:33.778532627 +0000 UTC m=+1295.733893683" watchObservedRunningTime="2025-12-04 18:00:33.781918386 +0000 UTC m=+1295.737279432" Dec 04 18:00:34 crc kubenswrapper[4733]: I1204 18:00:34.727146 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-rms44" event={"ID":"a9c17e0c-1bb9-4532-9af6-4e46aac1227e","Type":"ContainerStarted","Data":"f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c"} Dec 04 18:00:34 crc kubenswrapper[4733]: I1204 18:00:34.750684 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77f55878d5-rms44" podStartSLOduration=3.750669981 podStartE2EDuration="3.750669981s" podCreationTimestamp="2025-12-04 18:00:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:34.749059978 +0000 UTC m=+1296.704421044" watchObservedRunningTime="2025-12-04 18:00:34.750669981 +0000 UTC m=+1296.706031027" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.470899 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-84697bbbd9-9fltw"] Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.472664 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.474857 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.475065 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.507538 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84697bbbd9-9fltw"] Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.553649 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-public-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.553688 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-ovndb-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.553740 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-httpd-config\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.553822 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqsz6\" (UniqueName: \"kubernetes.io/projected/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-kube-api-access-hqsz6\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.553859 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-internal-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.554015 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-combined-ca-bundle\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.554218 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-config\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.616428 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.656067 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-httpd-config\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.656131 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqsz6\" (UniqueName: \"kubernetes.io/projected/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-kube-api-access-hqsz6\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.656169 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-internal-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.656208 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-combined-ca-bundle\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.656257 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-config\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.656279 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-public-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.656295 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-ovndb-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.661671 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-ovndb-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.662451 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-internal-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.662574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-httpd-config\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.665609 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-public-tls-certs\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.665712 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-combined-ca-bundle\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.666475 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-config\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.674258 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqsz6\" (UniqueName: \"kubernetes.io/projected/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-kube-api-access-hqsz6\") pod \"neutron-84697bbbd9-9fltw\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.697817 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.749924 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.750240 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.797370 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:35 crc kubenswrapper[4733]: I1204 18:00:35.884254 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 18:00:36 crc kubenswrapper[4733]: I1204 18:00:36.349754 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84697bbbd9-9fltw"] Dec 04 18:00:36 crc kubenswrapper[4733]: W1204 18:00:36.352741 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a9d9f43_f063_47fc_83b9_213c6ab06ef4.slice/crio-9bca0ad92f9667c2fd0009160b8cb4c642f40b2ed8fbf5b7ef8a90fb36e088b5 WatchSource:0}: Error finding container 9bca0ad92f9667c2fd0009160b8cb4c642f40b2ed8fbf5b7ef8a90fb36e088b5: Status 404 returned error can't find the container with id 9bca0ad92f9667c2fd0009160b8cb4c642f40b2ed8fbf5b7ef8a90fb36e088b5 Dec 04 18:00:36 crc kubenswrapper[4733]: I1204 18:00:36.757887 4733 generic.go:334] "Generic (PLEG): container finished" podID="abc26201-5065-4a19-9311-fc90d11401d0" containerID="d3243d7af0dfc453f4d69cd23ee6a59272dc78ba3c8ed8f8e372639c8284d8ae" exitCode=0 Dec 04 18:00:36 crc kubenswrapper[4733]: I1204 18:00:36.758767 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"abc26201-5065-4a19-9311-fc90d11401d0","Type":"ContainerDied","Data":"d3243d7af0dfc453f4d69cd23ee6a59272dc78ba3c8ed8f8e372639c8284d8ae"} Dec 04 18:00:36 crc kubenswrapper[4733]: I1204 18:00:36.760371 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84697bbbd9-9fltw" event={"ID":"4a9d9f43-f063-47fc-83b9-213c6ab06ef4","Type":"ContainerStarted","Data":"5a96bd3fd257dc89a4415f7d6d660060c72c788e7620add6273636ed06ed77b1"} Dec 04 18:00:36 crc kubenswrapper[4733]: I1204 18:00:36.760431 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84697bbbd9-9fltw" event={"ID":"4a9d9f43-f063-47fc-83b9-213c6ab06ef4","Type":"ContainerStarted","Data":"9bca0ad92f9667c2fd0009160b8cb4c642f40b2ed8fbf5b7ef8a90fb36e088b5"} Dec 04 18:00:36 crc kubenswrapper[4733]: I1204 18:00:36.773442 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.036061 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.083983 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-scripts\") pod \"abc26201-5065-4a19-9311-fc90d11401d0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.084358 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-log-httpd\") pod \"abc26201-5065-4a19-9311-fc90d11401d0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.084500 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88ssc\" (UniqueName: \"kubernetes.io/projected/abc26201-5065-4a19-9311-fc90d11401d0-kube-api-access-88ssc\") pod \"abc26201-5065-4a19-9311-fc90d11401d0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.084546 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-run-httpd\") pod \"abc26201-5065-4a19-9311-fc90d11401d0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.084570 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-combined-ca-bundle\") pod \"abc26201-5065-4a19-9311-fc90d11401d0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.084598 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-sg-core-conf-yaml\") pod \"abc26201-5065-4a19-9311-fc90d11401d0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.084665 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-config-data\") pod \"abc26201-5065-4a19-9311-fc90d11401d0\" (UID: \"abc26201-5065-4a19-9311-fc90d11401d0\") " Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.085461 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "abc26201-5065-4a19-9311-fc90d11401d0" (UID: "abc26201-5065-4a19-9311-fc90d11401d0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.085590 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "abc26201-5065-4a19-9311-fc90d11401d0" (UID: "abc26201-5065-4a19-9311-fc90d11401d0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.093381 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc26201-5065-4a19-9311-fc90d11401d0-kube-api-access-88ssc" (OuterVolumeSpecName: "kube-api-access-88ssc") pod "abc26201-5065-4a19-9311-fc90d11401d0" (UID: "abc26201-5065-4a19-9311-fc90d11401d0"). InnerVolumeSpecName "kube-api-access-88ssc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.098598 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-scripts" (OuterVolumeSpecName: "scripts") pod "abc26201-5065-4a19-9311-fc90d11401d0" (UID: "abc26201-5065-4a19-9311-fc90d11401d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.164316 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abc26201-5065-4a19-9311-fc90d11401d0" (UID: "abc26201-5065-4a19-9311-fc90d11401d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.186983 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.187011 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.187021 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.187028 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/abc26201-5065-4a19-9311-fc90d11401d0-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.187037 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88ssc\" (UniqueName: \"kubernetes.io/projected/abc26201-5065-4a19-9311-fc90d11401d0-kube-api-access-88ssc\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.189767 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-config-data" (OuterVolumeSpecName: "config-data") pod "abc26201-5065-4a19-9311-fc90d11401d0" (UID: "abc26201-5065-4a19-9311-fc90d11401d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.189874 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "abc26201-5065-4a19-9311-fc90d11401d0" (UID: "abc26201-5065-4a19-9311-fc90d11401d0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.288834 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.288866 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abc26201-5065-4a19-9311-fc90d11401d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.770630 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"abc26201-5065-4a19-9311-fc90d11401d0","Type":"ContainerDied","Data":"1dc2d491ae83567fd2aef13dccfba33cd1ee64e757d0786ae9951e67fd0e96f0"} Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.770687 4733 scope.go:117] "RemoveContainer" containerID="3894a7ff062a68f4a396cc9071b940d12d7b5f8e9e7fc8deb15155cc479476d6" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.770652 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.776938 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84697bbbd9-9fltw" event={"ID":"4a9d9f43-f063-47fc-83b9-213c6ab06ef4","Type":"ContainerStarted","Data":"04763b048c7378462084dba4301a110daf2dd03bc1d5722e8ceee83b2170fcc9"} Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.777827 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.782992 4733 generic.go:334] "Generic (PLEG): container finished" podID="8c53e91d-1256-408a-9b97-a048f425c441" containerID="9d57e43da79ab2c5d0b9a9159e21dc1548b4a16f3e0e5f81bcbb1076708a59e7" exitCode=0 Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.783073 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2khcb" event={"ID":"8c53e91d-1256-408a-9b97-a048f425c441","Type":"ContainerDied","Data":"9d57e43da79ab2c5d0b9a9159e21dc1548b4a16f3e0e5f81bcbb1076708a59e7"} Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.796662 4733 scope.go:117] "RemoveContainer" containerID="4a19da6f8268813a440ce518829aa523a1c4aa0a03e0e08d93c751d82e5bf4e0" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.827657 4733 scope.go:117] "RemoveContainer" containerID="d3243d7af0dfc453f4d69cd23ee6a59272dc78ba3c8ed8f8e372639c8284d8ae" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.838097 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-84697bbbd9-9fltw" podStartSLOduration=2.83807559 podStartE2EDuration="2.83807559s" podCreationTimestamp="2025-12-04 18:00:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:37.809433577 +0000 UTC m=+1299.764794653" watchObservedRunningTime="2025-12-04 18:00:37.83807559 +0000 UTC m=+1299.793436636" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.876730 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.890505 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.902427 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:00:37 crc kubenswrapper[4733]: E1204 18:00:37.902867 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="proxy-httpd" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.903240 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="proxy-httpd" Dec 04 18:00:37 crc kubenswrapper[4733]: E1204 18:00:37.903274 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="ceilometer-notification-agent" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.903284 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="ceilometer-notification-agent" Dec 04 18:00:37 crc kubenswrapper[4733]: E1204 18:00:37.903312 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="sg-core" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.903320 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="sg-core" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.903544 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="ceilometer-notification-agent" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.903570 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="sg-core" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.903593 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="abc26201-5065-4a19-9311-fc90d11401d0" containerName="proxy-httpd" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.905485 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.907890 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.908178 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 18:00:37 crc kubenswrapper[4733]: I1204 18:00:37.954997 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.003826 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-scripts\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.004150 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.004239 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-run-httpd\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.004272 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxnlt\" (UniqueName: \"kubernetes.io/projected/d6373063-26a1-4f42-9d95-79c54a11f8dc-kube-api-access-sxnlt\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.004321 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-config-data\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.004347 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.004448 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-log-httpd\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.106328 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-run-httpd\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.106365 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxnlt\" (UniqueName: \"kubernetes.io/projected/d6373063-26a1-4f42-9d95-79c54a11f8dc-kube-api-access-sxnlt\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.106404 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-config-data\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.106424 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.106503 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-log-httpd\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.106547 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-scripts\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.106567 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.108205 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-run-httpd\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.108236 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-log-httpd\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.112414 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.112781 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.112845 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-config-data\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.114175 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-scripts\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.127585 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxnlt\" (UniqueName: \"kubernetes.io/projected/d6373063-26a1-4f42-9d95-79c54a11f8dc-kube-api-access-sxnlt\") pod \"ceilometer-0\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.233234 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.373921 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abc26201-5065-4a19-9311-fc90d11401d0" path="/var/lib/kubelet/pods/abc26201-5065-4a19-9311-fc90d11401d0/volumes" Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.724460 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:00:38 crc kubenswrapper[4733]: W1204 18:00:38.764683 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6373063_26a1_4f42_9d95_79c54a11f8dc.slice/crio-089644e6f55e6c998efe4294a200d397d4ca51cd608492a5c720a1e760ae6d2a WatchSource:0}: Error finding container 089644e6f55e6c998efe4294a200d397d4ca51cd608492a5c720a1e760ae6d2a: Status 404 returned error can't find the container with id 089644e6f55e6c998efe4294a200d397d4ca51cd608492a5c720a1e760ae6d2a Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.796694 4733 generic.go:334] "Generic (PLEG): container finished" podID="3be601db-3698-4272-8005-083d5b93cbfd" containerID="e672d46632a950c5a99212c64ced6adecc83268c0542e4914c98073746bc1f0d" exitCode=0 Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.796749 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gbxkg" event={"ID":"3be601db-3698-4272-8005-083d5b93cbfd","Type":"ContainerDied","Data":"e672d46632a950c5a99212c64ced6adecc83268c0542e4914c98073746bc1f0d"} Dec 04 18:00:38 crc kubenswrapper[4733]: I1204 18:00:38.798820 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerStarted","Data":"089644e6f55e6c998efe4294a200d397d4ca51cd608492a5c720a1e760ae6d2a"} Dec 04 18:00:38 crc kubenswrapper[4733]: E1204 18:00:38.917614 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3be601db_3698_4272_8005_083d5b93cbfd.slice/crio-conmon-e672d46632a950c5a99212c64ced6adecc83268c0542e4914c98073746bc1f0d.scope\": RecentStats: unable to find data in memory cache]" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.363171 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2khcb" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.429551 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53e91d-1256-408a-9b97-a048f425c441-logs\") pod \"8c53e91d-1256-408a-9b97-a048f425c441\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.429620 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krgsx\" (UniqueName: \"kubernetes.io/projected/8c53e91d-1256-408a-9b97-a048f425c441-kube-api-access-krgsx\") pod \"8c53e91d-1256-408a-9b97-a048f425c441\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.429763 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-combined-ca-bundle\") pod \"8c53e91d-1256-408a-9b97-a048f425c441\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.429816 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-config-data\") pod \"8c53e91d-1256-408a-9b97-a048f425c441\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.429917 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-scripts\") pod \"8c53e91d-1256-408a-9b97-a048f425c441\" (UID: \"8c53e91d-1256-408a-9b97-a048f425c441\") " Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.430177 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c53e91d-1256-408a-9b97-a048f425c441-logs" (OuterVolumeSpecName: "logs") pod "8c53e91d-1256-408a-9b97-a048f425c441" (UID: "8c53e91d-1256-408a-9b97-a048f425c441"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.430603 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c53e91d-1256-408a-9b97-a048f425c441-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.434960 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c53e91d-1256-408a-9b97-a048f425c441-kube-api-access-krgsx" (OuterVolumeSpecName: "kube-api-access-krgsx") pod "8c53e91d-1256-408a-9b97-a048f425c441" (UID: "8c53e91d-1256-408a-9b97-a048f425c441"). InnerVolumeSpecName "kube-api-access-krgsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.440934 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-scripts" (OuterVolumeSpecName: "scripts") pod "8c53e91d-1256-408a-9b97-a048f425c441" (UID: "8c53e91d-1256-408a-9b97-a048f425c441"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.466566 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c53e91d-1256-408a-9b97-a048f425c441" (UID: "8c53e91d-1256-408a-9b97-a048f425c441"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.467371 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-config-data" (OuterVolumeSpecName: "config-data") pod "8c53e91d-1256-408a-9b97-a048f425c441" (UID: "8c53e91d-1256-408a-9b97-a048f425c441"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.531712 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krgsx\" (UniqueName: \"kubernetes.io/projected/8c53e91d-1256-408a-9b97-a048f425c441-kube-api-access-krgsx\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.531847 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.531912 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.531964 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c53e91d-1256-408a-9b97-a048f425c441-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.807433 4733 generic.go:334] "Generic (PLEG): container finished" podID="3bcc7f96-8d09-4d23-8e53-02b679d7851a" containerID="03be397f987be20236ee4d9b4181dbb81db3d0b00ead6b8a03cfb6abed63df74" exitCode=0 Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.807518 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-btknr" event={"ID":"3bcc7f96-8d09-4d23-8e53-02b679d7851a","Type":"ContainerDied","Data":"03be397f987be20236ee4d9b4181dbb81db3d0b00ead6b8a03cfb6abed63df74"} Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.811306 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2khcb" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.812284 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2khcb" event={"ID":"8c53e91d-1256-408a-9b97-a048f425c441","Type":"ContainerDied","Data":"b860c02be4704ea5232d214ba1d2e57c2c5fe25dff5df1b04e653a3d7f852572"} Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.812837 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b860c02be4704ea5232d214ba1d2e57c2c5fe25dff5df1b04e653a3d7f852572" Dec 04 18:00:39 crc kubenswrapper[4733]: I1204 18:00:39.813431 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerStarted","Data":"a9a9a9cdb9dcefd329ac3cb110ec4b2683940979f87de4e19d3577cc0127994b"} Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:39.999649 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d6df7889d-qkp4j"] Dec 04 18:00:40 crc kubenswrapper[4733]: E1204 18:00:39.999987 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c53e91d-1256-408a-9b97-a048f425c441" containerName="placement-db-sync" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:39.999998 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c53e91d-1256-408a-9b97-a048f425c441" containerName="placement-db-sync" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.000180 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c53e91d-1256-408a-9b97-a048f425c441" containerName="placement-db-sync" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.010246 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.013777 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.014271 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.014522 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vw67t" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.014679 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.014866 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.026557 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d6df7889d-qkp4j"] Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.039608 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-internal-tls-certs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.039651 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msvfc\" (UniqueName: \"kubernetes.io/projected/f575fafd-64f0-4401-82f1-f4a6a33b132e-kube-api-access-msvfc\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.039675 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f575fafd-64f0-4401-82f1-f4a6a33b132e-logs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.039704 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-combined-ca-bundle\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.039736 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-scripts\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.039755 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-config-data\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.039777 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-public-tls-certs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.143384 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-internal-tls-certs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.143819 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msvfc\" (UniqueName: \"kubernetes.io/projected/f575fafd-64f0-4401-82f1-f4a6a33b132e-kube-api-access-msvfc\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.143853 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f575fafd-64f0-4401-82f1-f4a6a33b132e-logs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.143892 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-combined-ca-bundle\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.143937 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-scripts\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.143962 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-config-data\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.143990 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-public-tls-certs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.144284 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f575fafd-64f0-4401-82f1-f4a6a33b132e-logs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.147745 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-public-tls-certs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.148118 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-config-data\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.152982 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-scripts\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.153151 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-internal-tls-certs\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.153403 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-combined-ca-bundle\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.164488 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msvfc\" (UniqueName: \"kubernetes.io/projected/f575fafd-64f0-4401-82f1-f4a6a33b132e-kube-api-access-msvfc\") pod \"placement-d6df7889d-qkp4j\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.201940 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.206123 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gbxkg" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.244611 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9spkp\" (UniqueName: \"kubernetes.io/projected/3be601db-3698-4272-8005-083d5b93cbfd-kube-api-access-9spkp\") pod \"3be601db-3698-4272-8005-083d5b93cbfd\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.244755 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-db-sync-config-data\") pod \"3be601db-3698-4272-8005-083d5b93cbfd\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.244888 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-combined-ca-bundle\") pod \"3be601db-3698-4272-8005-083d5b93cbfd\" (UID: \"3be601db-3698-4272-8005-083d5b93cbfd\") " Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.249278 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3be601db-3698-4272-8005-083d5b93cbfd" (UID: "3be601db-3698-4272-8005-083d5b93cbfd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.251997 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3be601db-3698-4272-8005-083d5b93cbfd-kube-api-access-9spkp" (OuterVolumeSpecName: "kube-api-access-9spkp") pod "3be601db-3698-4272-8005-083d5b93cbfd" (UID: "3be601db-3698-4272-8005-083d5b93cbfd"). InnerVolumeSpecName "kube-api-access-9spkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.273338 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3be601db-3698-4272-8005-083d5b93cbfd" (UID: "3be601db-3698-4272-8005-083d5b93cbfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.350001 4733 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.350043 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be601db-3698-4272-8005-083d5b93cbfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.350056 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9spkp\" (UniqueName: \"kubernetes.io/projected/3be601db-3698-4272-8005-083d5b93cbfd-kube-api-access-9spkp\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:40 crc kubenswrapper[4733]: W1204 18:00:40.692662 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf575fafd_64f0_4401_82f1_f4a6a33b132e.slice/crio-c9ac1c8691d89603fcbd0ce69f6273af7294c2c5fb8416393ad0228a2ed56f2b WatchSource:0}: Error finding container c9ac1c8691d89603fcbd0ce69f6273af7294c2c5fb8416393ad0228a2ed56f2b: Status 404 returned error can't find the container with id c9ac1c8691d89603fcbd0ce69f6273af7294c2c5fb8416393ad0228a2ed56f2b Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.693993 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d6df7889d-qkp4j"] Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.852325 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6df7889d-qkp4j" event={"ID":"f575fafd-64f0-4401-82f1-f4a6a33b132e","Type":"ContainerStarted","Data":"c9ac1c8691d89603fcbd0ce69f6273af7294c2c5fb8416393ad0228a2ed56f2b"} Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.854776 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerStarted","Data":"3d2f01ff68acc6abde2f03f1dfbcf4dbaf75f00f013effb036662df16bbc7102"} Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.870633 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gbxkg" Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.870852 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gbxkg" event={"ID":"3be601db-3698-4272-8005-083d5b93cbfd","Type":"ContainerDied","Data":"0723534f5e69a040f6f2edc098bea25ecbb6be9543d7abd95820e56edb8a3af4"} Dec 04 18:00:40 crc kubenswrapper[4733]: I1204 18:00:40.870895 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0723534f5e69a040f6f2edc098bea25ecbb6be9543d7abd95820e56edb8a3af4" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.141862 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-75746f8666-mw9nb"] Dec 04 18:00:41 crc kubenswrapper[4733]: E1204 18:00:41.142523 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3be601db-3698-4272-8005-083d5b93cbfd" containerName="barbican-db-sync" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.142541 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3be601db-3698-4272-8005-083d5b93cbfd" containerName="barbican-db-sync" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.142741 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3be601db-3698-4272-8005-083d5b93cbfd" containerName="barbican-db-sync" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.143754 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.149553 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jj8t9" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.149902 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.150027 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.160283 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-676cf9f69-6mwxl"] Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.162162 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.172849 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.172925 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnpnl\" (UniqueName: \"kubernetes.io/projected/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-kube-api-access-rnpnl\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.172958 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data-custom\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.172989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-combined-ca-bundle\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.173008 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-logs\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.173351 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.186787 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75746f8666-mw9nb"] Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.204836 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-676cf9f69-6mwxl"] Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.250988 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-rms44"] Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.251318 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77f55878d5-rms44" podUID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerName="dnsmasq-dns" containerID="cri-o://f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c" gracePeriod=10 Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.254162 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.273806 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-9n8jh"] Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.275407 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.279978 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280134 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280227 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnpnl\" (UniqueName: \"kubernetes.io/projected/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-kube-api-access-rnpnl\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data-custom\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280390 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-combined-ca-bundle\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-logs\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280666 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87f72b89-cea6-4243-8ec6-e1264b4901c3-logs\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280782 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z478s\" (UniqueName: \"kubernetes.io/projected/87f72b89-cea6-4243-8ec6-e1264b4901c3-kube-api-access-z478s\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280899 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-combined-ca-bundle\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.280967 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data-custom\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.283757 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-9n8jh"] Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.285504 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-logs\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.300340 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.300975 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-combined-ca-bundle\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.309428 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data-custom\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.314572 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnpnl\" (UniqueName: \"kubernetes.io/projected/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-kube-api-access-rnpnl\") pod \"barbican-keystone-listener-75746f8666-mw9nb\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.382943 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-svc\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383289 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-config\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383313 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383360 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnj8h\" (UniqueName: \"kubernetes.io/projected/7b60979a-506d-449f-b4d8-14cbbf6a77db-kube-api-access-tnj8h\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383395 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87f72b89-cea6-4243-8ec6-e1264b4901c3-logs\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383439 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z478s\" (UniqueName: \"kubernetes.io/projected/87f72b89-cea6-4243-8ec6-e1264b4901c3-kube-api-access-z478s\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383458 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-combined-ca-bundle\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383477 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data-custom\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383497 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383528 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.383568 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.385008 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87f72b89-cea6-4243-8ec6-e1264b4901c3-logs\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.391218 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-combined-ca-bundle\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.392721 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.393984 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data-custom\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.409326 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6d6b58894d-4kgnk"] Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.410710 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.413123 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.413215 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z478s\" (UniqueName: \"kubernetes.io/projected/87f72b89-cea6-4243-8ec6-e1264b4901c3-kube-api-access-z478s\") pod \"barbican-worker-676cf9f69-6mwxl\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.417188 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d6b58894d-4kgnk"] Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.465299 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-btknr" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487405 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487496 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487563 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqh86\" (UniqueName: \"kubernetes.io/projected/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-kube-api-access-kqh86\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487592 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-svc\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487615 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-config\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487635 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487659 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data-custom\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487696 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487727 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnj8h\" (UniqueName: \"kubernetes.io/projected/7b60979a-506d-449f-b4d8-14cbbf6a77db-kube-api-access-tnj8h\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487746 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-logs\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.487787 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-combined-ca-bundle\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.489445 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-swift-storage-0\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.490121 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-sb\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.491528 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-svc\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.491594 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-config\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.493326 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-nb\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.493497 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.502147 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.514686 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnj8h\" (UniqueName: \"kubernetes.io/projected/7b60979a-506d-449f-b4d8-14cbbf6a77db-kube-api-access-tnj8h\") pod \"dnsmasq-dns-687dbb56f-9n8jh\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.589532 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-db-sync-config-data\") pod \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.589613 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bcc7f96-8d09-4d23-8e53-02b679d7851a-etc-machine-id\") pod \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.589779 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkxzz\" (UniqueName: \"kubernetes.io/projected/3bcc7f96-8d09-4d23-8e53-02b679d7851a-kube-api-access-dkxzz\") pod \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.589863 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-config-data\") pod \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.589903 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-combined-ca-bundle\") pod \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.589967 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-scripts\") pod \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\" (UID: \"3bcc7f96-8d09-4d23-8e53-02b679d7851a\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.590218 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data-custom\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.590258 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.590291 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-logs\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.590338 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-combined-ca-bundle\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.590495 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqh86\" (UniqueName: \"kubernetes.io/projected/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-kube-api-access-kqh86\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.598484 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-logs\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.598552 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bcc7f96-8d09-4d23-8e53-02b679d7851a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3bcc7f96-8d09-4d23-8e53-02b679d7851a" (UID: "3bcc7f96-8d09-4d23-8e53-02b679d7851a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.601623 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data-custom\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.602134 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.606858 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcc7f96-8d09-4d23-8e53-02b679d7851a-kube-api-access-dkxzz" (OuterVolumeSpecName: "kube-api-access-dkxzz") pod "3bcc7f96-8d09-4d23-8e53-02b679d7851a" (UID: "3bcc7f96-8d09-4d23-8e53-02b679d7851a"). InnerVolumeSpecName "kube-api-access-dkxzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.608930 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-scripts" (OuterVolumeSpecName: "scripts") pod "3bcc7f96-8d09-4d23-8e53-02b679d7851a" (UID: "3bcc7f96-8d09-4d23-8e53-02b679d7851a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.612047 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3bcc7f96-8d09-4d23-8e53-02b679d7851a" (UID: "3bcc7f96-8d09-4d23-8e53-02b679d7851a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.615922 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-combined-ca-bundle\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.627365 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqh86\" (UniqueName: \"kubernetes.io/projected/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-kube-api-access-kqh86\") pod \"barbican-api-6d6b58894d-4kgnk\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.668214 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bcc7f96-8d09-4d23-8e53-02b679d7851a" (UID: "3bcc7f96-8d09-4d23-8e53-02b679d7851a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.668524 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.691921 4733 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.691954 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bcc7f96-8d09-4d23-8e53-02b679d7851a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.691963 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkxzz\" (UniqueName: \"kubernetes.io/projected/3bcc7f96-8d09-4d23-8e53-02b679d7851a-kube-api-access-dkxzz\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.691972 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.691982 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.712918 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-config-data" (OuterVolumeSpecName: "config-data") pod "3bcc7f96-8d09-4d23-8e53-02b679d7851a" (UID: "3bcc7f96-8d09-4d23-8e53-02b679d7851a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.741226 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.793935 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcc7f96-8d09-4d23-8e53-02b679d7851a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.874636 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.942899 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerStarted","Data":"b8d50c7b9a90929781ea8fa565d16c51cf9b07761341fe1fcc13b62cd52ebfa4"} Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.962628 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-btknr" event={"ID":"3bcc7f96-8d09-4d23-8e53-02b679d7851a","Type":"ContainerDied","Data":"fdbe17d093c9b8fe804454f2ab16a6d587d3fc47e85a0be922588a09409647cf"} Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.962665 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdbe17d093c9b8fe804454f2ab16a6d587d3fc47e85a0be922588a09409647cf" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.962732 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-btknr" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.986653 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6df7889d-qkp4j" event={"ID":"f575fafd-64f0-4401-82f1-f4a6a33b132e","Type":"ContainerStarted","Data":"d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8"} Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.986970 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6df7889d-qkp4j" event={"ID":"f575fafd-64f0-4401-82f1-f4a6a33b132e","Type":"ContainerStarted","Data":"606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2"} Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.988499 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.988552 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.999414 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-sb\") pod \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.999670 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz8jz\" (UniqueName: \"kubernetes.io/projected/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-kube-api-access-bz8jz\") pod \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.999700 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-config\") pod \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.999767 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-svc\") pod \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " Dec 04 18:00:41 crc kubenswrapper[4733]: I1204 18:00:41.999870 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-swift-storage-0\") pod \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:41.999908 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-nb\") pod \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\" (UID: \"a9c17e0c-1bb9-4532-9af6-4e46aac1227e\") " Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.021150 4733 generic.go:334] "Generic (PLEG): container finished" podID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerID="f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c" exitCode=0 Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.021195 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-rms44" event={"ID":"a9c17e0c-1bb9-4532-9af6-4e46aac1227e","Type":"ContainerDied","Data":"f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c"} Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.021218 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f55878d5-rms44" event={"ID":"a9c17e0c-1bb9-4532-9af6-4e46aac1227e","Type":"ContainerDied","Data":"07b2126c667bcfcd15c4bef866e469414ba38eeee30cc6b242cd9cfe4cb68983"} Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.021235 4733 scope.go:117] "RemoveContainer" containerID="f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.021385 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f55878d5-rms44" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.027022 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-kube-api-access-bz8jz" (OuterVolumeSpecName: "kube-api-access-bz8jz") pod "a9c17e0c-1bb9-4532-9af6-4e46aac1227e" (UID: "a9c17e0c-1bb9-4532-9af6-4e46aac1227e"). InnerVolumeSpecName "kube-api-access-bz8jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:42 crc kubenswrapper[4733]: W1204 18:00:42.041184 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1de9dcb_fad2_485d_b59b_14646f9ddfdd.slice/crio-d65a8d949cb2ae790623066d55bbd6b2c4db7377b065fcd19fe33a27fb04c78d WatchSource:0}: Error finding container d65a8d949cb2ae790623066d55bbd6b2c4db7377b065fcd19fe33a27fb04c78d: Status 404 returned error can't find the container with id d65a8d949cb2ae790623066d55bbd6b2c4db7377b065fcd19fe33a27fb04c78d Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.042951 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75746f8666-mw9nb"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.051446 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d6df7889d-qkp4j" podStartSLOduration=3.051409666 podStartE2EDuration="3.051409666s" podCreationTimestamp="2025-12-04 18:00:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:42.022240138 +0000 UTC m=+1303.977601184" watchObservedRunningTime="2025-12-04 18:00:42.051409666 +0000 UTC m=+1304.006770712" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.063004 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a9c17e0c-1bb9-4532-9af6-4e46aac1227e" (UID: "a9c17e0c-1bb9-4532-9af6-4e46aac1227e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.089211 4733 scope.go:117] "RemoveContainer" containerID="3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.101951 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.101985 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz8jz\" (UniqueName: \"kubernetes.io/projected/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-kube-api-access-bz8jz\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.104459 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a9c17e0c-1bb9-4532-9af6-4e46aac1227e" (UID: "a9c17e0c-1bb9-4532-9af6-4e46aac1227e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.142365 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a9c17e0c-1bb9-4532-9af6-4e46aac1227e" (UID: "a9c17e0c-1bb9-4532-9af6-4e46aac1227e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.143709 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-config" (OuterVolumeSpecName: "config") pod "a9c17e0c-1bb9-4532-9af6-4e46aac1227e" (UID: "a9c17e0c-1bb9-4532-9af6-4e46aac1227e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.165880 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:42 crc kubenswrapper[4733]: E1204 18:00:42.166336 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcc7f96-8d09-4d23-8e53-02b679d7851a" containerName="cinder-db-sync" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.166351 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcc7f96-8d09-4d23-8e53-02b679d7851a" containerName="cinder-db-sync" Dec 04 18:00:42 crc kubenswrapper[4733]: E1204 18:00:42.166366 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerName="init" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.166373 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerName="init" Dec 04 18:00:42 crc kubenswrapper[4733]: E1204 18:00:42.166403 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerName="dnsmasq-dns" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.166415 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerName="dnsmasq-dns" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.166623 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcc7f96-8d09-4d23-8e53-02b679d7851a" containerName="cinder-db-sync" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.166635 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerName="dnsmasq-dns" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.167724 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.171601 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.188736 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kh24l" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.188975 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a9c17e0c-1bb9-4532-9af6-4e46aac1227e" (UID: "a9c17e0c-1bb9-4532-9af6-4e46aac1227e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.189349 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.190248 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.190920 4733 scope.go:117] "RemoveContainer" containerID="f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c" Dec 04 18:00:42 crc kubenswrapper[4733]: E1204 18:00:42.193319 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c\": container with ID starting with f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c not found: ID does not exist" containerID="f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.193376 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c"} err="failed to get container status \"f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c\": rpc error: code = NotFound desc = could not find container \"f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c\": container with ID starting with f568188226c27ef99e2b9060ebf1b0e87528039eb50084436272c23cc944f75c not found: ID does not exist" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.193405 4733 scope.go:117] "RemoveContainer" containerID="3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b" Dec 04 18:00:42 crc kubenswrapper[4733]: E1204 18:00:42.197117 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b\": container with ID starting with 3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b not found: ID does not exist" containerID="3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.197175 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b"} err="failed to get container status \"3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b\": rpc error: code = NotFound desc = could not find container \"3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b\": container with ID starting with 3703f0071429e17fa65537df86187a32cd1e80a0059458749ee98a2403af963b not found: ID does not exist" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.207756 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b13bd75-d8e7-40d5-809e-0390df49cea6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.223009 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-scripts\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.223096 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.223147 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r8d2\" (UniqueName: \"kubernetes.io/projected/1b13bd75-d8e7-40d5-809e-0390df49cea6-kube-api-access-8r8d2\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.223209 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.223347 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.228135 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.228158 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.228172 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.228182 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9c17e0c-1bb9-4532-9af6-4e46aac1227e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.234779 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-9n8jh"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.288404 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.302837 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-922zq"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.304502 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.312873 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-676cf9f69-6mwxl"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.319719 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-922zq"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.329454 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b13bd75-d8e7-40d5-809e-0390df49cea6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.329508 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-scripts\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.329555 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.329583 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r8d2\" (UniqueName: \"kubernetes.io/projected/1b13bd75-d8e7-40d5-809e-0390df49cea6-kube-api-access-8r8d2\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.329600 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b13bd75-d8e7-40d5-809e-0390df49cea6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.329614 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.331171 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.333930 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.334296 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.333674 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.338960 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.339997 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-scripts\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.342030 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.351426 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.354528 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r8d2\" (UniqueName: \"kubernetes.io/projected/1b13bd75-d8e7-40d5-809e-0390df49cea6-kube-api-access-8r8d2\") pod \"cinder-scheduler-0\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.365026 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.386241 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-rms44"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.392530 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77f55878d5-rms44"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.436743 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7373f67d-6afa-43a8-881c-973d33b1db9e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.436806 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8dw2\" (UniqueName: \"kubernetes.io/projected/7373f67d-6afa-43a8-881c-973d33b1db9e-kube-api-access-w8dw2\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.436839 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-config\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.436893 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-scripts\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.436933 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data-custom\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.436950 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.437010 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7373f67d-6afa-43a8-881c-973d33b1db9e-logs\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.437031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv9x2\" (UniqueName: \"kubernetes.io/projected/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-kube-api-access-lv9x2\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.437054 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.437073 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.437102 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.437142 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.437169 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.468562 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-9n8jh"] Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.519919 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7373f67d-6afa-43a8-881c-973d33b1db9e-logs\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539116 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv9x2\" (UniqueName: \"kubernetes.io/projected/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-kube-api-access-lv9x2\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539143 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539164 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539214 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539244 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539265 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539285 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7373f67d-6afa-43a8-881c-973d33b1db9e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539311 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8dw2\" (UniqueName: \"kubernetes.io/projected/7373f67d-6afa-43a8-881c-973d33b1db9e-kube-api-access-w8dw2\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539330 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-config\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539354 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-scripts\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539375 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data-custom\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.539391 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.540183 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.540526 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7373f67d-6afa-43a8-881c-973d33b1db9e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.540781 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-svc\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.541114 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7373f67d-6afa-43a8-881c-973d33b1db9e-logs\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.543238 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.544412 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-config\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.550492 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data-custom\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.551221 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-scripts\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.544908 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-swift-storage-0\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.558858 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.564838 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8dw2\" (UniqueName: \"kubernetes.io/projected/7373f67d-6afa-43a8-881c-973d33b1db9e-kube-api-access-w8dw2\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.567088 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv9x2\" (UniqueName: \"kubernetes.io/projected/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-kube-api-access-lv9x2\") pod \"dnsmasq-dns-77d8c9c7-922zq\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.587111 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.623487 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.660601 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 18:00:42 crc kubenswrapper[4733]: I1204 18:00:42.712941 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d6b58894d-4kgnk"] Dec 04 18:00:42 crc kubenswrapper[4733]: W1204 18:00:42.853960 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbee8dcc9_2257_4d48_9b20_ce7c2e475ace.slice/crio-790fb762696011c4bf3b6dcab8bc4b34fbb161f15d7760497d2acdcbad87b671 WatchSource:0}: Error finding container 790fb762696011c4bf3b6dcab8bc4b34fbb161f15d7760497d2acdcbad87b671: Status 404 returned error can't find the container with id 790fb762696011c4bf3b6dcab8bc4b34fbb161f15d7760497d2acdcbad87b671 Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.038409 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" event={"ID":"c1de9dcb-fad2-485d-b59b-14646f9ddfdd","Type":"ContainerStarted","Data":"d65a8d949cb2ae790623066d55bbd6b2c4db7377b065fcd19fe33a27fb04c78d"} Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.043009 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d6b58894d-4kgnk" event={"ID":"bee8dcc9-2257-4d48-9b20-ce7c2e475ace","Type":"ContainerStarted","Data":"790fb762696011c4bf3b6dcab8bc4b34fbb161f15d7760497d2acdcbad87b671"} Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.046417 4733 generic.go:334] "Generic (PLEG): container finished" podID="7b60979a-506d-449f-b4d8-14cbbf6a77db" containerID="6ceefc365ead64012eafe67e5f1a32dcdf1a0b4c92e752b0602deb0ad965b386" exitCode=0 Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.046471 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" event={"ID":"7b60979a-506d-449f-b4d8-14cbbf6a77db","Type":"ContainerDied","Data":"6ceefc365ead64012eafe67e5f1a32dcdf1a0b4c92e752b0602deb0ad965b386"} Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.046488 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" event={"ID":"7b60979a-506d-449f-b4d8-14cbbf6a77db","Type":"ContainerStarted","Data":"fcee31094575c64c7042046988e9b724c3edbc544ad4591675fa78e569b65cd4"} Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.060155 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerStarted","Data":"b8dd0239fe51987da1c26aa092d25f2f708220112e9e888bc7a2dc790b0ac8bb"} Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.060813 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.064168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-676cf9f69-6mwxl" event={"ID":"87f72b89-cea6-4243-8ec6-e1264b4901c3","Type":"ContainerStarted","Data":"ff5801d26152910887f23748ab78ea20d11f11f76e866284f98bf5c2f9f5b66a"} Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.092310 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.497355072 podStartE2EDuration="6.092294677s" podCreationTimestamp="2025-12-04 18:00:37 +0000 UTC" firstStartedPulling="2025-12-04 18:00:38.772479621 +0000 UTC m=+1300.727840657" lastFinishedPulling="2025-12-04 18:00:42.367419216 +0000 UTC m=+1304.322780262" observedRunningTime="2025-12-04 18:00:43.090050419 +0000 UTC m=+1305.045411465" watchObservedRunningTime="2025-12-04 18:00:43.092294677 +0000 UTC m=+1305.047655723" Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.170888 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.314689 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-922zq"] Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.434379 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:43 crc kubenswrapper[4733]: W1204 18:00:43.769585 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7373f67d_6afa_43a8_881c_973d33b1db9e.slice/crio-a225464ea4bbc70162d567ae207d65e29ead51ceaa13a14bbc0f674313ad68e3 WatchSource:0}: Error finding container a225464ea4bbc70162d567ae207d65e29ead51ceaa13a14bbc0f674313ad68e3: Status 404 returned error can't find the container with id a225464ea4bbc70162d567ae207d65e29ead51ceaa13a14bbc0f674313ad68e3 Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.816773 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.973820 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-config\") pod \"7b60979a-506d-449f-b4d8-14cbbf6a77db\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.973927 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-svc\") pod \"7b60979a-506d-449f-b4d8-14cbbf6a77db\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.973957 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-sb\") pod \"7b60979a-506d-449f-b4d8-14cbbf6a77db\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.973990 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnj8h\" (UniqueName: \"kubernetes.io/projected/7b60979a-506d-449f-b4d8-14cbbf6a77db-kube-api-access-tnj8h\") pod \"7b60979a-506d-449f-b4d8-14cbbf6a77db\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.974052 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-swift-storage-0\") pod \"7b60979a-506d-449f-b4d8-14cbbf6a77db\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.974068 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-nb\") pod \"7b60979a-506d-449f-b4d8-14cbbf6a77db\" (UID: \"7b60979a-506d-449f-b4d8-14cbbf6a77db\") " Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.978726 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b60979a-506d-449f-b4d8-14cbbf6a77db-kube-api-access-tnj8h" (OuterVolumeSpecName: "kube-api-access-tnj8h") pod "7b60979a-506d-449f-b4d8-14cbbf6a77db" (UID: "7b60979a-506d-449f-b4d8-14cbbf6a77db"). InnerVolumeSpecName "kube-api-access-tnj8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.998229 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7b60979a-506d-449f-b4d8-14cbbf6a77db" (UID: "7b60979a-506d-449f-b4d8-14cbbf6a77db"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:43 crc kubenswrapper[4733]: I1204 18:00:43.999026 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7b60979a-506d-449f-b4d8-14cbbf6a77db" (UID: "7b60979a-506d-449f-b4d8-14cbbf6a77db"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.002315 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7b60979a-506d-449f-b4d8-14cbbf6a77db" (UID: "7b60979a-506d-449f-b4d8-14cbbf6a77db"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.002893 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-config" (OuterVolumeSpecName: "config") pod "7b60979a-506d-449f-b4d8-14cbbf6a77db" (UID: "7b60979a-506d-449f-b4d8-14cbbf6a77db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.005328 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7b60979a-506d-449f-b4d8-14cbbf6a77db" (UID: "7b60979a-506d-449f-b4d8-14cbbf6a77db"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.074035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" event={"ID":"dffa9d54-372c-4a62-8955-f0ae3ae2f89a","Type":"ContainerStarted","Data":"7aef1254ec57ad7dc144e5aa1863c799c9a9e0c0376da4ce9612ed438d7308b3"} Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.076344 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.076453 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.076524 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.076592 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.076688 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b60979a-506d-449f-b4d8-14cbbf6a77db-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.076767 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnj8h\" (UniqueName: \"kubernetes.io/projected/7b60979a-506d-449f-b4d8-14cbbf6a77db-kube-api-access-tnj8h\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.078629 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7373f67d-6afa-43a8-881c-973d33b1db9e","Type":"ContainerStarted","Data":"a225464ea4bbc70162d567ae207d65e29ead51ceaa13a14bbc0f674313ad68e3"} Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.080852 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d6b58894d-4kgnk" event={"ID":"bee8dcc9-2257-4d48-9b20-ce7c2e475ace","Type":"ContainerStarted","Data":"23c7156715dc312cf9c68d957a663cf2b3309212fd9c27c4155a016b0acf19e4"} Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.080992 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d6b58894d-4kgnk" event={"ID":"bee8dcc9-2257-4d48-9b20-ce7c2e475ace","Type":"ContainerStarted","Data":"7d9eae6ba40801169aa6ce00e726bdaf1409a1ccdc06c38b31af84be68313c6a"} Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.081121 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.081473 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.083605 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" event={"ID":"7b60979a-506d-449f-b4d8-14cbbf6a77db","Type":"ContainerDied","Data":"fcee31094575c64c7042046988e9b724c3edbc544ad4591675fa78e569b65cd4"} Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.083727 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687dbb56f-9n8jh" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.083733 4733 scope.go:117] "RemoveContainer" containerID="6ceefc365ead64012eafe67e5f1a32dcdf1a0b4c92e752b0602deb0ad965b386" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.085703 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b13bd75-d8e7-40d5-809e-0390df49cea6","Type":"ContainerStarted","Data":"569ede874dc181e0ed528f58c19b1e524466a242a4d6b5d224548bf01e55ebd6"} Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.194029 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6d6b58894d-4kgnk" podStartSLOduration=3.194008149 podStartE2EDuration="3.194008149s" podCreationTimestamp="2025-12-04 18:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:44.105226994 +0000 UTC m=+1306.060588050" watchObservedRunningTime="2025-12-04 18:00:44.194008149 +0000 UTC m=+1306.149369195" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.253412 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-9n8jh"] Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.266561 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-687dbb56f-9n8jh"] Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.362878 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b60979a-506d-449f-b4d8-14cbbf6a77db" path="/var/lib/kubelet/pods/7b60979a-506d-449f-b4d8-14cbbf6a77db/volumes" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.363943 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" path="/var/lib/kubelet/pods/a9c17e0c-1bb9-4532-9af6-4e46aac1227e/volumes" Dec 04 18:00:44 crc kubenswrapper[4733]: I1204 18:00:44.480778 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.110296 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7373f67d-6afa-43a8-881c-973d33b1db9e","Type":"ContainerStarted","Data":"ae36ec866154ca2849b515df94657cfd959d268145f097c6f6bc10a94a0b9d5a"} Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.113005 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" event={"ID":"c1de9dcb-fad2-485d-b59b-14646f9ddfdd","Type":"ContainerStarted","Data":"52c25be5cf28de218adebb584f7e6fe433e56d68d1b191f31c36c67cb6d7c1a1"} Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.113068 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" event={"ID":"c1de9dcb-fad2-485d-b59b-14646f9ddfdd","Type":"ContainerStarted","Data":"b2b953dbdaed812d6f01b224d3af9906a0d0da35f8c6eaeed44b254649426640"} Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.118350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b13bd75-d8e7-40d5-809e-0390df49cea6","Type":"ContainerStarted","Data":"b4d8f3153bf28ff3a9f828d772e2a07d32adc19bd7e0e719a18409eed1f32ea2"} Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.124265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-676cf9f69-6mwxl" event={"ID":"87f72b89-cea6-4243-8ec6-e1264b4901c3","Type":"ContainerStarted","Data":"594133e5106de2f14d436d478b5b6fa76670e3bcd1bb710de01b002b98c8b79a"} Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.124310 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-676cf9f69-6mwxl" event={"ID":"87f72b89-cea6-4243-8ec6-e1264b4901c3","Type":"ContainerStarted","Data":"a6c4a59c75c37a61feea92a8e8114a64759adb31de0de0bbb5f0d5725e5b60bd"} Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.126948 4733 generic.go:334] "Generic (PLEG): container finished" podID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" containerID="138f9c8aa9a6459369e3d5b34cfe94fe9cdb0f0de5f46db942505f32795f02cf" exitCode=0 Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.128277 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" event={"ID":"dffa9d54-372c-4a62-8955-f0ae3ae2f89a","Type":"ContainerDied","Data":"138f9c8aa9a6459369e3d5b34cfe94fe9cdb0f0de5f46db942505f32795f02cf"} Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.154306 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" podStartSLOduration=1.937212709 podStartE2EDuration="4.15428052s" podCreationTimestamp="2025-12-04 18:00:41 +0000 UTC" firstStartedPulling="2025-12-04 18:00:42.063294238 +0000 UTC m=+1304.018655284" lastFinishedPulling="2025-12-04 18:00:44.280362049 +0000 UTC m=+1306.235723095" observedRunningTime="2025-12-04 18:00:45.137320334 +0000 UTC m=+1307.092681380" watchObservedRunningTime="2025-12-04 18:00:45.15428052 +0000 UTC m=+1307.109641566" Dec 04 18:00:45 crc kubenswrapper[4733]: I1204 18:00:45.188783 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-676cf9f69-6mwxl" podStartSLOduration=2.178510317 podStartE2EDuration="4.188764018s" podCreationTimestamp="2025-12-04 18:00:41 +0000 UTC" firstStartedPulling="2025-12-04 18:00:42.270018485 +0000 UTC m=+1304.225379531" lastFinishedPulling="2025-12-04 18:00:44.280272186 +0000 UTC m=+1306.235633232" observedRunningTime="2025-12-04 18:00:45.183103649 +0000 UTC m=+1307.138464695" watchObservedRunningTime="2025-12-04 18:00:45.188764018 +0000 UTC m=+1307.144125064" Dec 04 18:00:46 crc kubenswrapper[4733]: I1204 18:00:46.138706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" event={"ID":"dffa9d54-372c-4a62-8955-f0ae3ae2f89a","Type":"ContainerStarted","Data":"b83a809aa02d2009fa0815c6ef37a6f8e9afc105fbbb1fbd2aaba3ca86931a93"} Dec 04 18:00:46 crc kubenswrapper[4733]: I1204 18:00:46.139068 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:46 crc kubenswrapper[4733]: I1204 18:00:46.140737 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7373f67d-6afa-43a8-881c-973d33b1db9e","Type":"ContainerStarted","Data":"955bd0b5e2ca5b306180b82c116dccf5ef469929cc9228bccc842381d191fd9f"} Dec 04 18:00:46 crc kubenswrapper[4733]: I1204 18:00:46.141241 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerName="cinder-api-log" containerID="cri-o://ae36ec866154ca2849b515df94657cfd959d268145f097c6f6bc10a94a0b9d5a" gracePeriod=30 Dec 04 18:00:46 crc kubenswrapper[4733]: I1204 18:00:46.141263 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerName="cinder-api" containerID="cri-o://955bd0b5e2ca5b306180b82c116dccf5ef469929cc9228bccc842381d191fd9f" gracePeriod=30 Dec 04 18:00:46 crc kubenswrapper[4733]: I1204 18:00:46.178338 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" podStartSLOduration=4.178302719 podStartE2EDuration="4.178302719s" podCreationTimestamp="2025-12-04 18:00:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:46.175933846 +0000 UTC m=+1308.131294892" watchObservedRunningTime="2025-12-04 18:00:46.178302719 +0000 UTC m=+1308.133663765" Dec 04 18:00:46 crc kubenswrapper[4733]: I1204 18:00:46.213113 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.213090253 podStartE2EDuration="4.213090253s" podCreationTimestamp="2025-12-04 18:00:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:46.208574865 +0000 UTC m=+1308.163935911" watchObservedRunningTime="2025-12-04 18:00:46.213090253 +0000 UTC m=+1308.168451329" Dec 04 18:00:46 crc kubenswrapper[4733]: I1204 18:00:46.597195 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77f55878d5-rms44" podUID="a9c17e0c-1bb9-4532-9af6-4e46aac1227e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.153:5353: i/o timeout" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.175091 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b13bd75-d8e7-40d5-809e-0390df49cea6","Type":"ContainerStarted","Data":"17effa55d5014bbb37b70198b0ff6c9472bb2d684386773a2411cee6b42323de"} Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.192495 4733 generic.go:334] "Generic (PLEG): container finished" podID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerID="955bd0b5e2ca5b306180b82c116dccf5ef469929cc9228bccc842381d191fd9f" exitCode=0 Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.192535 4733 generic.go:334] "Generic (PLEG): container finished" podID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerID="ae36ec866154ca2849b515df94657cfd959d268145f097c6f6bc10a94a0b9d5a" exitCode=143 Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.192903 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7373f67d-6afa-43a8-881c-973d33b1db9e","Type":"ContainerDied","Data":"955bd0b5e2ca5b306180b82c116dccf5ef469929cc9228bccc842381d191fd9f"} Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.192947 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7373f67d-6afa-43a8-881c-973d33b1db9e","Type":"ContainerDied","Data":"ae36ec866154ca2849b515df94657cfd959d268145f097c6f6bc10a94a0b9d5a"} Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.207143 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.102721732 podStartE2EDuration="5.207124983s" podCreationTimestamp="2025-12-04 18:00:42 +0000 UTC" firstStartedPulling="2025-12-04 18:00:43.187144172 +0000 UTC m=+1305.142505218" lastFinishedPulling="2025-12-04 18:00:44.291547423 +0000 UTC m=+1306.246908469" observedRunningTime="2025-12-04 18:00:47.199694978 +0000 UTC m=+1309.155056024" watchObservedRunningTime="2025-12-04 18:00:47.207124983 +0000 UTC m=+1309.162486019" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.255883 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.362403 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8dw2\" (UniqueName: \"kubernetes.io/projected/7373f67d-6afa-43a8-881c-973d33b1db9e-kube-api-access-w8dw2\") pod \"7373f67d-6afa-43a8-881c-973d33b1db9e\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.362726 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-scripts\") pod \"7373f67d-6afa-43a8-881c-973d33b1db9e\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.362881 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7373f67d-6afa-43a8-881c-973d33b1db9e-etc-machine-id\") pod \"7373f67d-6afa-43a8-881c-973d33b1db9e\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.363028 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-combined-ca-bundle\") pod \"7373f67d-6afa-43a8-881c-973d33b1db9e\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.363497 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data\") pod \"7373f67d-6afa-43a8-881c-973d33b1db9e\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.363610 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7373f67d-6afa-43a8-881c-973d33b1db9e-logs\") pod \"7373f67d-6afa-43a8-881c-973d33b1db9e\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.363694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data-custom\") pod \"7373f67d-6afa-43a8-881c-973d33b1db9e\" (UID: \"7373f67d-6afa-43a8-881c-973d33b1db9e\") " Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.363032 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7373f67d-6afa-43a8-881c-973d33b1db9e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7373f67d-6afa-43a8-881c-973d33b1db9e" (UID: "7373f67d-6afa-43a8-881c-973d33b1db9e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.364161 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7373f67d-6afa-43a8-881c-973d33b1db9e-logs" (OuterVolumeSpecName: "logs") pod "7373f67d-6afa-43a8-881c-973d33b1db9e" (UID: "7373f67d-6afa-43a8-881c-973d33b1db9e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.364666 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7373f67d-6afa-43a8-881c-973d33b1db9e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.366014 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7373f67d-6afa-43a8-881c-973d33b1db9e-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.369225 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7373f67d-6afa-43a8-881c-973d33b1db9e-kube-api-access-w8dw2" (OuterVolumeSpecName: "kube-api-access-w8dw2") pod "7373f67d-6afa-43a8-881c-973d33b1db9e" (UID: "7373f67d-6afa-43a8-881c-973d33b1db9e"). InnerVolumeSpecName "kube-api-access-w8dw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.375321 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-scripts" (OuterVolumeSpecName: "scripts") pod "7373f67d-6afa-43a8-881c-973d33b1db9e" (UID: "7373f67d-6afa-43a8-881c-973d33b1db9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.376954 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7373f67d-6afa-43a8-881c-973d33b1db9e" (UID: "7373f67d-6afa-43a8-881c-973d33b1db9e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.392967 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7373f67d-6afa-43a8-881c-973d33b1db9e" (UID: "7373f67d-6afa-43a8-881c-973d33b1db9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.431952 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data" (OuterVolumeSpecName: "config-data") pod "7373f67d-6afa-43a8-881c-973d33b1db9e" (UID: "7373f67d-6afa-43a8-881c-973d33b1db9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.468360 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.468707 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.468724 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.468736 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7373f67d-6afa-43a8-881c-973d33b1db9e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.468744 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8dw2\" (UniqueName: \"kubernetes.io/projected/7373f67d-6afa-43a8-881c-973d33b1db9e-kube-api-access-w8dw2\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.521451 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.969255 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7f757c77dd-bzp5x"] Dec 04 18:00:47 crc kubenswrapper[4733]: E1204 18:00:47.969767 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerName="cinder-api" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.969789 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerName="cinder-api" Dec 04 18:00:47 crc kubenswrapper[4733]: E1204 18:00:47.969841 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b60979a-506d-449f-b4d8-14cbbf6a77db" containerName="init" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.969851 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b60979a-506d-449f-b4d8-14cbbf6a77db" containerName="init" Dec 04 18:00:47 crc kubenswrapper[4733]: E1204 18:00:47.969866 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerName="cinder-api-log" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.969875 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerName="cinder-api-log" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.970108 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b60979a-506d-449f-b4d8-14cbbf6a77db" containerName="init" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.970139 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerName="cinder-api" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.970153 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" containerName="cinder-api-log" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.971331 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.974405 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.977054 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 04 18:00:47 crc kubenswrapper[4733]: I1204 18:00:47.990997 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f757c77dd-bzp5x"] Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.079694 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-internal-tls-certs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.079753 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-combined-ca-bundle\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.079779 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.079912 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-public-tls-certs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.079956 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52st8\" (UniqueName: \"kubernetes.io/projected/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-kube-api-access-52st8\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.080021 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-logs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.080060 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data-custom\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.182061 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-combined-ca-bundle\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.182128 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.182175 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-public-tls-certs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.182208 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52st8\" (UniqueName: \"kubernetes.io/projected/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-kube-api-access-52st8\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.182248 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-logs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.182284 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data-custom\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.182355 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-internal-tls-certs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.183065 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-logs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.187407 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-internal-tls-certs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.188118 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-combined-ca-bundle\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.188621 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.191748 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data-custom\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.193574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-public-tls-certs\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.206787 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.206785 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7373f67d-6afa-43a8-881c-973d33b1db9e","Type":"ContainerDied","Data":"a225464ea4bbc70162d567ae207d65e29ead51ceaa13a14bbc0f674313ad68e3"} Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.207074 4733 scope.go:117] "RemoveContainer" containerID="955bd0b5e2ca5b306180b82c116dccf5ef469929cc9228bccc842381d191fd9f" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.208612 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52st8\" (UniqueName: \"kubernetes.io/projected/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-kube-api-access-52st8\") pod \"barbican-api-7f757c77dd-bzp5x\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.291246 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.303783 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.311871 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.318366 4733 scope.go:117] "RemoveContainer" containerID="ae36ec866154ca2849b515df94657cfd959d268145f097c6f6bc10a94a0b9d5a" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.319574 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.321536 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.329291 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.329516 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.329672 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.365596 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7373f67d-6afa-43a8-881c-973d33b1db9e" path="/var/lib/kubelet/pods/7373f67d-6afa-43a8-881c-973d33b1db9e/volumes" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.366328 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492353 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492414 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-scripts\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492440 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12b58db1-9574-4081-a3c1-2853f003fa8f-logs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492484 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492507 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492550 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492618 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dtf5\" (UniqueName: \"kubernetes.io/projected/12b58db1-9574-4081-a3c1-2853f003fa8f-kube-api-access-8dtf5\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492665 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.492684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12b58db1-9574-4081-a3c1-2853f003fa8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594082 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594121 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12b58db1-9574-4081-a3c1-2853f003fa8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594149 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594194 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-scripts\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594219 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12b58db1-9574-4081-a3c1-2853f003fa8f-logs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594261 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594257 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12b58db1-9574-4081-a3c1-2853f003fa8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594282 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594480 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594577 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dtf5\" (UniqueName: \"kubernetes.io/projected/12b58db1-9574-4081-a3c1-2853f003fa8f-kube-api-access-8dtf5\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.594705 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12b58db1-9574-4081-a3c1-2853f003fa8f-logs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.599215 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-scripts\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.599374 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.600110 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.602577 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.605331 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.608231 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.613911 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dtf5\" (UniqueName: \"kubernetes.io/projected/12b58db1-9574-4081-a3c1-2853f003fa8f-kube-api-access-8dtf5\") pod \"cinder-api-0\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.740384 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 18:00:48 crc kubenswrapper[4733]: I1204 18:00:48.851663 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f757c77dd-bzp5x"] Dec 04 18:00:49 crc kubenswrapper[4733]: I1204 18:00:49.190474 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:00:49 crc kubenswrapper[4733]: W1204 18:00:49.195980 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12b58db1_9574_4081_a3c1_2853f003fa8f.slice/crio-3545c73604d8f3cb46724ea7074e847de530750900a13c82a111f09584041895 WatchSource:0}: Error finding container 3545c73604d8f3cb46724ea7074e847de530750900a13c82a111f09584041895: Status 404 returned error can't find the container with id 3545c73604d8f3cb46724ea7074e847de530750900a13c82a111f09584041895 Dec 04 18:00:49 crc kubenswrapper[4733]: I1204 18:00:49.233469 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12b58db1-9574-4081-a3c1-2853f003fa8f","Type":"ContainerStarted","Data":"3545c73604d8f3cb46724ea7074e847de530750900a13c82a111f09584041895"} Dec 04 18:00:49 crc kubenswrapper[4733]: I1204 18:00:49.237647 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f757c77dd-bzp5x" event={"ID":"3c5f9b26-6c76-4af7-a811-d7d763f74ed6","Type":"ContainerStarted","Data":"1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d"} Dec 04 18:00:49 crc kubenswrapper[4733]: I1204 18:00:49.237699 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f757c77dd-bzp5x" event={"ID":"3c5f9b26-6c76-4af7-a811-d7d763f74ed6","Type":"ContainerStarted","Data":"d7f944db0d8b755d5a29260a295bbc66eb68e4af34fcc983a42e2ddc19c795cb"} Dec 04 18:00:50 crc kubenswrapper[4733]: I1204 18:00:50.249936 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12b58db1-9574-4081-a3c1-2853f003fa8f","Type":"ContainerStarted","Data":"d366e224d098b7e5af2be8bd08ebc779d71f9239fa811ef64c92bd18e26bace0"} Dec 04 18:00:50 crc kubenswrapper[4733]: I1204 18:00:50.251687 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f757c77dd-bzp5x" event={"ID":"3c5f9b26-6c76-4af7-a811-d7d763f74ed6","Type":"ContainerStarted","Data":"a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd"} Dec 04 18:00:50 crc kubenswrapper[4733]: I1204 18:00:50.251989 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:50 crc kubenswrapper[4733]: I1204 18:00:50.252026 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:00:50 crc kubenswrapper[4733]: I1204 18:00:50.273784 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7f757c77dd-bzp5x" podStartSLOduration=3.273765515 podStartE2EDuration="3.273765515s" podCreationTimestamp="2025-12-04 18:00:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:50.269114263 +0000 UTC m=+1312.224475309" watchObservedRunningTime="2025-12-04 18:00:50.273765515 +0000 UTC m=+1312.229126561" Dec 04 18:00:51 crc kubenswrapper[4733]: I1204 18:00:51.261451 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12b58db1-9574-4081-a3c1-2853f003fa8f","Type":"ContainerStarted","Data":"5cd076e1396593ece03d523ffdfe9f53cc72e35677e8aab21b5111c2bcce8fa1"} Dec 04 18:00:51 crc kubenswrapper[4733]: I1204 18:00:51.261831 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 18:00:51 crc kubenswrapper[4733]: I1204 18:00:51.290033 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.29001194 podStartE2EDuration="3.29001194s" podCreationTimestamp="2025-12-04 18:00:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:00:51.283358334 +0000 UTC m=+1313.238719420" watchObservedRunningTime="2025-12-04 18:00:51.29001194 +0000 UTC m=+1313.245372996" Dec 04 18:00:52 crc kubenswrapper[4733]: I1204 18:00:52.625950 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:00:52 crc kubenswrapper[4733]: I1204 18:00:52.712059 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-lpxxd"] Dec 04 18:00:52 crc kubenswrapper[4733]: I1204 18:00:52.712326 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" podUID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" containerName="dnsmasq-dns" containerID="cri-o://099a0dd66081b1742a35d04c3412573bb6ed9b1b5b14e8add80a8d0e79aa25ab" gracePeriod=10 Dec 04 18:00:52 crc kubenswrapper[4733]: I1204 18:00:52.832761 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 18:00:52 crc kubenswrapper[4733]: I1204 18:00:52.902643 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.285657 4733 generic.go:334] "Generic (PLEG): container finished" podID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" containerID="099a0dd66081b1742a35d04c3412573bb6ed9b1b5b14e8add80a8d0e79aa25ab" exitCode=0 Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.285913 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerName="cinder-scheduler" containerID="cri-o://b4d8f3153bf28ff3a9f828d772e2a07d32adc19bd7e0e719a18409eed1f32ea2" gracePeriod=30 Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.286326 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" event={"ID":"c20a96e3-9c22-48fc-ba00-4f1a13350ecd","Type":"ContainerDied","Data":"099a0dd66081b1742a35d04c3412573bb6ed9b1b5b14e8add80a8d0e79aa25ab"} Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.286378 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" event={"ID":"c20a96e3-9c22-48fc-ba00-4f1a13350ecd","Type":"ContainerDied","Data":"8f264d3b1793a9e00a2e15c65efd8f18f68d8e57af3b40ae1dbd8c0e5793a658"} Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.286388 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f264d3b1793a9e00a2e15c65efd8f18f68d8e57af3b40ae1dbd8c0e5793a658" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.286724 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerName="probe" containerID="cri-o://17effa55d5014bbb37b70198b0ff6c9472bb2d684386773a2411cee6b42323de" gracePeriod=30 Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.303689 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.319863 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.410343 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.503065 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-nb\") pod \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.503336 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-sb\") pod \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.503474 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-config\") pod \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.503616 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-swift-storage-0\") pod \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.503705 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-svc\") pod \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.503776 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzfrt\" (UniqueName: \"kubernetes.io/projected/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-kube-api-access-mzfrt\") pod \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\" (UID: \"c20a96e3-9c22-48fc-ba00-4f1a13350ecd\") " Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.548041 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-kube-api-access-mzfrt" (OuterVolumeSpecName: "kube-api-access-mzfrt") pod "c20a96e3-9c22-48fc-ba00-4f1a13350ecd" (UID: "c20a96e3-9c22-48fc-ba00-4f1a13350ecd"). InnerVolumeSpecName "kube-api-access-mzfrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.605474 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-config" (OuterVolumeSpecName: "config") pod "c20a96e3-9c22-48fc-ba00-4f1a13350ecd" (UID: "c20a96e3-9c22-48fc-ba00-4f1a13350ecd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.608518 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.608573 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzfrt\" (UniqueName: \"kubernetes.io/projected/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-kube-api-access-mzfrt\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.609310 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c20a96e3-9c22-48fc-ba00-4f1a13350ecd" (UID: "c20a96e3-9c22-48fc-ba00-4f1a13350ecd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.613477 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c20a96e3-9c22-48fc-ba00-4f1a13350ecd" (UID: "c20a96e3-9c22-48fc-ba00-4f1a13350ecd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.615752 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c20a96e3-9c22-48fc-ba00-4f1a13350ecd" (UID: "c20a96e3-9c22-48fc-ba00-4f1a13350ecd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.637838 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c20a96e3-9c22-48fc-ba00-4f1a13350ecd" (UID: "c20a96e3-9c22-48fc-ba00-4f1a13350ecd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.710720 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.710776 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.710787 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:53 crc kubenswrapper[4733]: I1204 18:00:53.710812 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c20a96e3-9c22-48fc-ba00-4f1a13350ecd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:54 crc kubenswrapper[4733]: I1204 18:00:54.297002 4733 generic.go:334] "Generic (PLEG): container finished" podID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerID="17effa55d5014bbb37b70198b0ff6c9472bb2d684386773a2411cee6b42323de" exitCode=0 Dec 04 18:00:54 crc kubenswrapper[4733]: I1204 18:00:54.297981 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b13bd75-d8e7-40d5-809e-0390df49cea6","Type":"ContainerDied","Data":"17effa55d5014bbb37b70198b0ff6c9472bb2d684386773a2411cee6b42323de"} Dec 04 18:00:54 crc kubenswrapper[4733]: I1204 18:00:54.298219 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7884648fd9-lpxxd" Dec 04 18:00:54 crc kubenswrapper[4733]: I1204 18:00:54.332654 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-lpxxd"] Dec 04 18:00:54 crc kubenswrapper[4733]: I1204 18:00:54.366983 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7884648fd9-lpxxd"] Dec 04 18:00:55 crc kubenswrapper[4733]: I1204 18:00:55.134761 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.325019 4733 generic.go:334] "Generic (PLEG): container finished" podID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerID="b4d8f3153bf28ff3a9f828d772e2a07d32adc19bd7e0e719a18409eed1f32ea2" exitCode=0 Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.325196 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b13bd75-d8e7-40d5-809e-0390df49cea6","Type":"ContainerDied","Data":"b4d8f3153bf28ff3a9f828d772e2a07d32adc19bd7e0e719a18409eed1f32ea2"} Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.326385 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b13bd75-d8e7-40d5-809e-0390df49cea6","Type":"ContainerDied","Data":"569ede874dc181e0ed528f58c19b1e524466a242a4d6b5d224548bf01e55ebd6"} Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.326488 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="569ede874dc181e0ed528f58c19b1e524466a242a4d6b5d224548bf01e55ebd6" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.349670 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" path="/var/lib/kubelet/pods/c20a96e3-9c22-48fc-ba00-4f1a13350ecd/volumes" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.379090 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.458299 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-combined-ca-bundle\") pod \"1b13bd75-d8e7-40d5-809e-0390df49cea6\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.458348 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b13bd75-d8e7-40d5-809e-0390df49cea6-etc-machine-id\") pod \"1b13bd75-d8e7-40d5-809e-0390df49cea6\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.458434 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data\") pod \"1b13bd75-d8e7-40d5-809e-0390df49cea6\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.458488 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b13bd75-d8e7-40d5-809e-0390df49cea6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1b13bd75-d8e7-40d5-809e-0390df49cea6" (UID: "1b13bd75-d8e7-40d5-809e-0390df49cea6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.458516 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-scripts\") pod \"1b13bd75-d8e7-40d5-809e-0390df49cea6\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.458564 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r8d2\" (UniqueName: \"kubernetes.io/projected/1b13bd75-d8e7-40d5-809e-0390df49cea6-kube-api-access-8r8d2\") pod \"1b13bd75-d8e7-40d5-809e-0390df49cea6\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.458593 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data-custom\") pod \"1b13bd75-d8e7-40d5-809e-0390df49cea6\" (UID: \"1b13bd75-d8e7-40d5-809e-0390df49cea6\") " Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.458943 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b13bd75-d8e7-40d5-809e-0390df49cea6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.482161 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b13bd75-d8e7-40d5-809e-0390df49cea6-kube-api-access-8r8d2" (OuterVolumeSpecName: "kube-api-access-8r8d2") pod "1b13bd75-d8e7-40d5-809e-0390df49cea6" (UID: "1b13bd75-d8e7-40d5-809e-0390df49cea6"). InnerVolumeSpecName "kube-api-access-8r8d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.482204 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1b13bd75-d8e7-40d5-809e-0390df49cea6" (UID: "1b13bd75-d8e7-40d5-809e-0390df49cea6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.484899 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-scripts" (OuterVolumeSpecName: "scripts") pod "1b13bd75-d8e7-40d5-809e-0390df49cea6" (UID: "1b13bd75-d8e7-40d5-809e-0390df49cea6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.530646 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b13bd75-d8e7-40d5-809e-0390df49cea6" (UID: "1b13bd75-d8e7-40d5-809e-0390df49cea6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.563548 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.563703 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r8d2\" (UniqueName: \"kubernetes.io/projected/1b13bd75-d8e7-40d5-809e-0390df49cea6-kube-api-access-8r8d2\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.563783 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.563877 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.581940 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data" (OuterVolumeSpecName: "config-data") pod "1b13bd75-d8e7-40d5-809e-0390df49cea6" (UID: "1b13bd75-d8e7-40d5-809e-0390df49cea6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:00:56 crc kubenswrapper[4733]: I1204 18:00:56.666293 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b13bd75-d8e7-40d5-809e-0390df49cea6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.335371 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.375734 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.388022 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.400695 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:57 crc kubenswrapper[4733]: E1204 18:00:57.401161 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerName="cinder-scheduler" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.401180 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerName="cinder-scheduler" Dec 04 18:00:57 crc kubenswrapper[4733]: E1204 18:00:57.401203 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" containerName="init" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.401210 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" containerName="init" Dec 04 18:00:57 crc kubenswrapper[4733]: E1204 18:00:57.401221 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" containerName="dnsmasq-dns" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.401227 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" containerName="dnsmasq-dns" Dec 04 18:00:57 crc kubenswrapper[4733]: E1204 18:00:57.401256 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerName="probe" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.401262 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerName="probe" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.401421 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c20a96e3-9c22-48fc-ba00-4f1a13350ecd" containerName="dnsmasq-dns" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.401440 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerName="cinder-scheduler" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.401455 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" containerName="probe" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.402379 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.404582 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.414935 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.480601 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.480921 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.481043 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.481162 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7df4d1e-71e7-4c40-a3e4-27b484799d09-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.481291 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj8fc\" (UniqueName: \"kubernetes.io/projected/f7df4d1e-71e7-4c40-a3e4-27b484799d09-kube-api-access-qj8fc\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.481502 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-scripts\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.583411 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-scripts\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.583513 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.583558 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.583584 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.583615 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7df4d1e-71e7-4c40-a3e4-27b484799d09-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.583632 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj8fc\" (UniqueName: \"kubernetes.io/projected/f7df4d1e-71e7-4c40-a3e4-27b484799d09-kube-api-access-qj8fc\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.584010 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7df4d1e-71e7-4c40-a3e4-27b484799d09-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.588635 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.588689 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.592225 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-scripts\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.592715 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.610936 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj8fc\" (UniqueName: \"kubernetes.io/projected/f7df4d1e-71e7-4c40-a3e4-27b484799d09-kube-api-access-qj8fc\") pod \"cinder-scheduler-0\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " pod="openstack/cinder-scheduler-0" Dec 04 18:00:57 crc kubenswrapper[4733]: I1204 18:00:57.724087 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 18:00:58 crc kubenswrapper[4733]: W1204 18:00:58.168851 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7df4d1e_71e7_4c40_a3e4_27b484799d09.slice/crio-496c76265d05efa68b400a73f274d401dcb2d8c058678bef6ad325d66387d6a8 WatchSource:0}: Error finding container 496c76265d05efa68b400a73f274d401dcb2d8c058678bef6ad325d66387d6a8: Status 404 returned error can't find the container with id 496c76265d05efa68b400a73f274d401dcb2d8c058678bef6ad325d66387d6a8 Dec 04 18:00:58 crc kubenswrapper[4733]: I1204 18:00:58.171590 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:00:58 crc kubenswrapper[4733]: I1204 18:00:58.378840 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b13bd75-d8e7-40d5-809e-0390df49cea6" path="/var/lib/kubelet/pods/1b13bd75-d8e7-40d5-809e-0390df49cea6/volumes" Dec 04 18:00:58 crc kubenswrapper[4733]: I1204 18:00:58.380151 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f7df4d1e-71e7-4c40-a3e4-27b484799d09","Type":"ContainerStarted","Data":"496c76265d05efa68b400a73f274d401dcb2d8c058678bef6ad325d66387d6a8"} Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.049111 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.050433 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.055012 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.055064 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-gfnjl" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.055182 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.083843 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.213158 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config-secret\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.213305 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk2q8\" (UniqueName: \"kubernetes.io/projected/7b19ede9-4e23-416d-bc7a-3674ed551c21-kube-api-access-zk2q8\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.213393 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.213422 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.315349 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.315610 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.315632 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config-secret\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.315741 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk2q8\" (UniqueName: \"kubernetes.io/projected/7b19ede9-4e23-416d-bc7a-3674ed551c21-kube-api-access-zk2q8\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.317400 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.327439 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.330720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config-secret\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.350098 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk2q8\" (UniqueName: \"kubernetes.io/projected/7b19ede9-4e23-416d-bc7a-3674ed551c21-kube-api-access-zk2q8\") pod \"openstackclient\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " pod="openstack/openstackclient" Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.395689 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f7df4d1e-71e7-4c40-a3e4-27b484799d09","Type":"ContainerStarted","Data":"9aa9ac8aec81ac8172b07681bf3447f72d402bc17fafe55ed65c2f79900cd3fb"} Dec 04 18:00:59 crc kubenswrapper[4733]: I1204 18:00:59.470893 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.001344 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.020421 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.089675 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.158948 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29414521-gw72h"] Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.160450 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.175893 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6d6b58894d-4kgnk"] Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.176132 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6d6b58894d-4kgnk" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api-log" containerID="cri-o://7d9eae6ba40801169aa6ce00e726bdaf1409a1ccdc06c38b31af84be68313c6a" gracePeriod=30 Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.176252 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6d6b58894d-4kgnk" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api" containerID="cri-o://23c7156715dc312cf9c68d957a663cf2b3309212fd9c27c4155a016b0acf19e4" gracePeriod=30 Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.197925 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414521-gw72h"] Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.233615 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-fernet-keys\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.234271 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpb4s\" (UniqueName: \"kubernetes.io/projected/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-kube-api-access-hpb4s\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.234420 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-combined-ca-bundle\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.234618 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-config-data\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.346096 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpb4s\" (UniqueName: \"kubernetes.io/projected/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-kube-api-access-hpb4s\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.346253 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-combined-ca-bundle\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.346526 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-config-data\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.346601 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-fernet-keys\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.366499 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-fernet-keys\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.367430 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-config-data\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.367692 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-combined-ca-bundle\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.384689 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpb4s\" (UniqueName: \"kubernetes.io/projected/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-kube-api-access-hpb4s\") pod \"keystone-cron-29414521-gw72h\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.413204 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f7df4d1e-71e7-4c40-a3e4-27b484799d09","Type":"ContainerStarted","Data":"3044b3de56fc15c0557352a235e182ec8eaffbf427385fd0dad2161e82e28cd2"} Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.431728 4733 generic.go:334] "Generic (PLEG): container finished" podID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerID="7d9eae6ba40801169aa6ce00e726bdaf1409a1ccdc06c38b31af84be68313c6a" exitCode=143 Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.431803 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d6b58894d-4kgnk" event={"ID":"bee8dcc9-2257-4d48-9b20-ce7c2e475ace","Type":"ContainerDied","Data":"7d9eae6ba40801169aa6ce00e726bdaf1409a1ccdc06c38b31af84be68313c6a"} Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.441356 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"7b19ede9-4e23-416d-bc7a-3674ed551c21","Type":"ContainerStarted","Data":"dd2552a52e606992fd0d93f83f3c9eacf2686b424ed0dbe674142bb023634cdc"} Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.454612 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.454595428 podStartE2EDuration="3.454595428s" podCreationTimestamp="2025-12-04 18:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:00.452118062 +0000 UTC m=+1322.407479108" watchObservedRunningTime="2025-12-04 18:01:00.454595428 +0000 UTC m=+1322.409956484" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.499171 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:00 crc kubenswrapper[4733]: I1204 18:01:00.965594 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414521-gw72h"] Dec 04 18:01:00 crc kubenswrapper[4733]: W1204 18:01:00.974486 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95c361e7_6b27_4624_8a71_bccd6e6c0cd0.slice/crio-d8ea883ab14a1102d8179a6ef47cc42ced90c93c26d46e78a53ac11b1050a2f3 WatchSource:0}: Error finding container d8ea883ab14a1102d8179a6ef47cc42ced90c93c26d46e78a53ac11b1050a2f3: Status 404 returned error can't find the container with id d8ea883ab14a1102d8179a6ef47cc42ced90c93c26d46e78a53ac11b1050a2f3 Dec 04 18:01:01 crc kubenswrapper[4733]: I1204 18:01:01.343647 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 18:01:01 crc kubenswrapper[4733]: I1204 18:01:01.508085 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414521-gw72h" event={"ID":"95c361e7-6b27-4624-8a71-bccd6e6c0cd0","Type":"ContainerStarted","Data":"04ca6ac5aad6dfaafa91564d2deaf7c701fa148811b46f58283eb9c5dd6a73d0"} Dec 04 18:01:01 crc kubenswrapper[4733]: I1204 18:01:01.508119 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414521-gw72h" event={"ID":"95c361e7-6b27-4624-8a71-bccd6e6c0cd0","Type":"ContainerStarted","Data":"d8ea883ab14a1102d8179a6ef47cc42ced90c93c26d46e78a53ac11b1050a2f3"} Dec 04 18:01:01 crc kubenswrapper[4733]: I1204 18:01:01.842723 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:01:01 crc kubenswrapper[4733]: I1204 18:01:01.868322 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29414521-gw72h" podStartSLOduration=1.868298904 podStartE2EDuration="1.868298904s" podCreationTimestamp="2025-12-04 18:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:01.530365647 +0000 UTC m=+1323.485726693" watchObservedRunningTime="2025-12-04 18:01:01.868298904 +0000 UTC m=+1323.823659950" Dec 04 18:01:02 crc kubenswrapper[4733]: I1204 18:01:02.727921 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.347137 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d6b58894d-4kgnk" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:34234->10.217.0.161:9311: read: connection reset by peer" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.347205 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d6b58894d-4kgnk" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:34248->10.217.0.161:9311: read: connection reset by peer" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.540656 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-59945f4865-9zsr4"] Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.542875 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.545864 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.546133 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.546333 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.553321 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-59945f4865-9zsr4"] Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.564493 4733 generic.go:334] "Generic (PLEG): container finished" podID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerID="23c7156715dc312cf9c68d957a663cf2b3309212fd9c27c4155a016b0acf19e4" exitCode=0 Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.564528 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d6b58894d-4kgnk" event={"ID":"bee8dcc9-2257-4d48-9b20-ce7c2e475ace","Type":"ContainerDied","Data":"23c7156715dc312cf9c68d957a663cf2b3309212fd9c27c4155a016b0acf19e4"} Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.573159 4733 generic.go:334] "Generic (PLEG): container finished" podID="95c361e7-6b27-4624-8a71-bccd6e6c0cd0" containerID="04ca6ac5aad6dfaafa91564d2deaf7c701fa148811b46f58283eb9c5dd6a73d0" exitCode=0 Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.573211 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414521-gw72h" event={"ID":"95c361e7-6b27-4624-8a71-bccd6e6c0cd0","Type":"ContainerDied","Data":"04ca6ac5aad6dfaafa91564d2deaf7c701fa148811b46f58283eb9c5dd6a73d0"} Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.627171 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-run-httpd\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.627228 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-public-tls-certs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.627275 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-config-data\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.627425 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-etc-swift\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.627479 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-log-httpd\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.627504 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-combined-ca-bundle\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.627528 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkmgs\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-kube-api-access-xkmgs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.627555 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-internal-tls-certs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.729489 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-log-httpd\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.729538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-combined-ca-bundle\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.729557 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkmgs\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-kube-api-access-xkmgs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.729582 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-internal-tls-certs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.729629 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-run-httpd\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.729652 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-public-tls-certs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.729686 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-config-data\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.729731 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-etc-swift\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.730247 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-log-httpd\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.730553 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-run-httpd\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.735416 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-etc-swift\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.737307 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-internal-tls-certs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.737410 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-combined-ca-bundle\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.737970 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-public-tls-certs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.738499 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-config-data\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.746905 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkmgs\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-kube-api-access-xkmgs\") pod \"swift-proxy-59945f4865-9zsr4\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.820511 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.874110 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.931952 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data\") pod \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.932051 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqh86\" (UniqueName: \"kubernetes.io/projected/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-kube-api-access-kqh86\") pod \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.932079 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data-custom\") pod \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.932114 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-combined-ca-bundle\") pod \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.932261 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-logs\") pod \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\" (UID: \"bee8dcc9-2257-4d48-9b20-ce7c2e475ace\") " Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.932750 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-logs" (OuterVolumeSpecName: "logs") pod "bee8dcc9-2257-4d48-9b20-ce7c2e475ace" (UID: "bee8dcc9-2257-4d48-9b20-ce7c2e475ace"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.935714 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-kube-api-access-kqh86" (OuterVolumeSpecName: "kube-api-access-kqh86") pod "bee8dcc9-2257-4d48-9b20-ce7c2e475ace" (UID: "bee8dcc9-2257-4d48-9b20-ce7c2e475ace"). InnerVolumeSpecName "kube-api-access-kqh86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.936305 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bee8dcc9-2257-4d48-9b20-ce7c2e475ace" (UID: "bee8dcc9-2257-4d48-9b20-ce7c2e475ace"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.981878 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bee8dcc9-2257-4d48-9b20-ce7c2e475ace" (UID: "bee8dcc9-2257-4d48-9b20-ce7c2e475ace"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:03 crc kubenswrapper[4733]: I1204 18:01:03.993890 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data" (OuterVolumeSpecName: "config-data") pod "bee8dcc9-2257-4d48-9b20-ce7c2e475ace" (UID: "bee8dcc9-2257-4d48-9b20-ce7c2e475ace"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.034510 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.034537 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.034548 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqh86\" (UniqueName: \"kubernetes.io/projected/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-kube-api-access-kqh86\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.034559 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.034567 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee8dcc9-2257-4d48-9b20-ce7c2e475ace-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.384233 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-59945f4865-9zsr4"] Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.565126 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.566068 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="ceilometer-central-agent" containerID="cri-o://a9a9a9cdb9dcefd329ac3cb110ec4b2683940979f87de4e19d3577cc0127994b" gracePeriod=30 Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.567365 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="sg-core" containerID="cri-o://b8d50c7b9a90929781ea8fa565d16c51cf9b07761341fe1fcc13b62cd52ebfa4" gracePeriod=30 Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.567518 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="proxy-httpd" containerID="cri-o://b8dd0239fe51987da1c26aa092d25f2f708220112e9e888bc7a2dc790b0ac8bb" gracePeriod=30 Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.567631 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="ceilometer-notification-agent" containerID="cri-o://3d2f01ff68acc6abde2f03f1dfbcf4dbaf75f00f013effb036662df16bbc7102" gracePeriod=30 Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.575442 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.156:3000/\": EOF" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.587182 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59945f4865-9zsr4" event={"ID":"d78c8335-c2e4-4c49-8b77-98fec3cab751","Type":"ContainerStarted","Data":"0cd6a42aad9bf69767b8f6ddd75025503a74ca5623b580aa89c5758a7b6ce2ec"} Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.590178 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d6b58894d-4kgnk" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.591054 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d6b58894d-4kgnk" event={"ID":"bee8dcc9-2257-4d48-9b20-ce7c2e475ace","Type":"ContainerDied","Data":"790fb762696011c4bf3b6dcab8bc4b34fbb161f15d7760497d2acdcbad87b671"} Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.591097 4733 scope.go:117] "RemoveContainer" containerID="23c7156715dc312cf9c68d957a663cf2b3309212fd9c27c4155a016b0acf19e4" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.630347 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6d6b58894d-4kgnk"] Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.638706 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6d6b58894d-4kgnk"] Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.644623 4733 scope.go:117] "RemoveContainer" containerID="7d9eae6ba40801169aa6ce00e726bdaf1409a1ccdc06c38b31af84be68313c6a" Dec 04 18:01:04 crc kubenswrapper[4733]: I1204 18:01:04.929774 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.055231 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-fernet-keys\") pod \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.055304 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-config-data\") pod \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.055374 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpb4s\" (UniqueName: \"kubernetes.io/projected/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-kube-api-access-hpb4s\") pod \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.055461 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-combined-ca-bundle\") pod \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\" (UID: \"95c361e7-6b27-4624-8a71-bccd6e6c0cd0\") " Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.061948 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "95c361e7-6b27-4624-8a71-bccd6e6c0cd0" (UID: "95c361e7-6b27-4624-8a71-bccd6e6c0cd0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.062028 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-kube-api-access-hpb4s" (OuterVolumeSpecName: "kube-api-access-hpb4s") pod "95c361e7-6b27-4624-8a71-bccd6e6c0cd0" (UID: "95c361e7-6b27-4624-8a71-bccd6e6c0cd0"). InnerVolumeSpecName "kube-api-access-hpb4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.089207 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95c361e7-6b27-4624-8a71-bccd6e6c0cd0" (UID: "95c361e7-6b27-4624-8a71-bccd6e6c0cd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.123937 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-config-data" (OuterVolumeSpecName: "config-data") pod "95c361e7-6b27-4624-8a71-bccd6e6c0cd0" (UID: "95c361e7-6b27-4624-8a71-bccd6e6c0cd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.158151 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.158202 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.158214 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpb4s\" (UniqueName: \"kubernetes.io/projected/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-kube-api-access-hpb4s\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.158223 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c361e7-6b27-4624-8a71-bccd6e6c0cd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.602021 4733 generic.go:334] "Generic (PLEG): container finished" podID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerID="b8dd0239fe51987da1c26aa092d25f2f708220112e9e888bc7a2dc790b0ac8bb" exitCode=0 Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.602342 4733 generic.go:334] "Generic (PLEG): container finished" podID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerID="b8d50c7b9a90929781ea8fa565d16c51cf9b07761341fe1fcc13b62cd52ebfa4" exitCode=2 Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.602356 4733 generic.go:334] "Generic (PLEG): container finished" podID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerID="a9a9a9cdb9dcefd329ac3cb110ec4b2683940979f87de4e19d3577cc0127994b" exitCode=0 Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.602106 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerDied","Data":"b8dd0239fe51987da1c26aa092d25f2f708220112e9e888bc7a2dc790b0ac8bb"} Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.602455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerDied","Data":"b8d50c7b9a90929781ea8fa565d16c51cf9b07761341fe1fcc13b62cd52ebfa4"} Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.602509 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerDied","Data":"a9a9a9cdb9dcefd329ac3cb110ec4b2683940979f87de4e19d3577cc0127994b"} Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.606498 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414521-gw72h" event={"ID":"95c361e7-6b27-4624-8a71-bccd6e6c0cd0","Type":"ContainerDied","Data":"d8ea883ab14a1102d8179a6ef47cc42ced90c93c26d46e78a53ac11b1050a2f3"} Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.606527 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414521-gw72h" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.606531 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8ea883ab14a1102d8179a6ef47cc42ced90c93c26d46e78a53ac11b1050a2f3" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.609277 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59945f4865-9zsr4" event={"ID":"d78c8335-c2e4-4c49-8b77-98fec3cab751","Type":"ContainerStarted","Data":"f130a979912194eb23dc33f38388f755ef2b29dc69498aa6ea5c028aff72030b"} Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.609438 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.609467 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.609479 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59945f4865-9zsr4" event={"ID":"d78c8335-c2e4-4c49-8b77-98fec3cab751","Type":"ContainerStarted","Data":"7a464410f79ac8c4ad9cc078a988fa37fc20c807e7742c3550f9f6de916082bb"} Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.655406 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-59945f4865-9zsr4" podStartSLOduration=2.655386022 podStartE2EDuration="2.655386022s" podCreationTimestamp="2025-12-04 18:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:05.633765393 +0000 UTC m=+1327.589126439" watchObservedRunningTime="2025-12-04 18:01:05.655386022 +0000 UTC m=+1327.610747068" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.811469 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.872009 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-98f7dbcd8-vdtg8"] Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.872244 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-98f7dbcd8-vdtg8" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerName="neutron-api" containerID="cri-o://262ab0c4956d796d1019c1aab9759043a7cd8f1e971e1c279de42fb81935131d" gracePeriod=30 Dec 04 18:01:05 crc kubenswrapper[4733]: I1204 18:01:05.872370 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-98f7dbcd8-vdtg8" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerName="neutron-httpd" containerID="cri-o://5ae8831a27bc1dd563aa1309ef400056f32ce6d82dc70a522d2f89456bde35d9" gracePeriod=30 Dec 04 18:01:06 crc kubenswrapper[4733]: I1204 18:01:06.366988 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" path="/var/lib/kubelet/pods/bee8dcc9-2257-4d48-9b20-ce7c2e475ace/volumes" Dec 04 18:01:06 crc kubenswrapper[4733]: I1204 18:01:06.621148 4733 generic.go:334] "Generic (PLEG): container finished" podID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerID="5ae8831a27bc1dd563aa1309ef400056f32ce6d82dc70a522d2f89456bde35d9" exitCode=0 Dec 04 18:01:06 crc kubenswrapper[4733]: I1204 18:01:06.621254 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-98f7dbcd8-vdtg8" event={"ID":"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab","Type":"ContainerDied","Data":"5ae8831a27bc1dd563aa1309ef400056f32ce6d82dc70a522d2f89456bde35d9"} Dec 04 18:01:07 crc kubenswrapper[4733]: I1204 18:01:07.631206 4733 generic.go:334] "Generic (PLEG): container finished" podID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerID="3d2f01ff68acc6abde2f03f1dfbcf4dbaf75f00f013effb036662df16bbc7102" exitCode=0 Dec 04 18:01:07 crc kubenswrapper[4733]: I1204 18:01:07.631253 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerDied","Data":"3d2f01ff68acc6abde2f03f1dfbcf4dbaf75f00f013effb036662df16bbc7102"} Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.138722 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.234149 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.156:3000/\": dial tcp 10.217.0.156:3000: connect: connection refused" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.794341 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.800975 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" containerName="glance-log" containerID="cri-o://db6f3a34d675cc354963a482ff62b85e52c39c97b79d0c7f54bfdaf0383617c0" gracePeriod=30 Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.801456 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" containerName="glance-httpd" containerID="cri-o://ec9864790923d6c96b26352e7b30f7d1edab920785e6bd453aae8436fb4d67ab" gracePeriod=30 Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.904659 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-vk5ks"] Dec 04 18:01:08 crc kubenswrapper[4733]: E1204 18:01:08.905077 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.905098 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api" Dec 04 18:01:08 crc kubenswrapper[4733]: E1204 18:01:08.905114 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95c361e7-6b27-4624-8a71-bccd6e6c0cd0" containerName="keystone-cron" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.905123 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="95c361e7-6b27-4624-8a71-bccd6e6c0cd0" containerName="keystone-cron" Dec 04 18:01:08 crc kubenswrapper[4733]: E1204 18:01:08.905135 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api-log" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.905142 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api-log" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.905324 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api-log" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.905352 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="95c361e7-6b27-4624-8a71-bccd6e6c0cd0" containerName="keystone-cron" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.905369 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bee8dcc9-2257-4d48-9b20-ce7c2e475ace" containerName="barbican-api" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.913492 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.920938 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vk5ks"] Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.961307 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5c0f-account-create-update-2gwwc"] Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.962430 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.965199 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 04 18:01:08 crc kubenswrapper[4733]: I1204 18:01:08.969499 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5c0f-account-create-update-2gwwc"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.045349 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bbdf249-90f9-4685-9418-1ecd4729ac09-operator-scripts\") pod \"nova-api-db-create-vk5ks\" (UID: \"9bbdf249-90f9-4685-9418-1ecd4729ac09\") " pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.045969 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m56mx\" (UniqueName: \"kubernetes.io/projected/9bbdf249-90f9-4685-9418-1ecd4729ac09-kube-api-access-m56mx\") pod \"nova-api-db-create-vk5ks\" (UID: \"9bbdf249-90f9-4685-9418-1ecd4729ac09\") " pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.053636 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-r28fb"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.055166 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.080186 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-r28fb"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.127934 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-6ac2-account-create-update-l2v8g"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.129568 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.133953 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.135083 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-ztq89"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.136161 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.142771 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6ac2-account-create-update-l2v8g"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.149813 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56tmh\" (UniqueName: \"kubernetes.io/projected/e8436db7-0c0c-4c64-aea2-2266437d9d44-kube-api-access-56tmh\") pod \"nova-api-5c0f-account-create-update-2gwwc\" (UID: \"e8436db7-0c0c-4c64-aea2-2266437d9d44\") " pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.149885 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bbdf249-90f9-4685-9418-1ecd4729ac09-operator-scripts\") pod \"nova-api-db-create-vk5ks\" (UID: \"9bbdf249-90f9-4685-9418-1ecd4729ac09\") " pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.149931 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8436db7-0c0c-4c64-aea2-2266437d9d44-operator-scripts\") pod \"nova-api-5c0f-account-create-update-2gwwc\" (UID: \"e8436db7-0c0c-4c64-aea2-2266437d9d44\") " pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.149987 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m56mx\" (UniqueName: \"kubernetes.io/projected/9bbdf249-90f9-4685-9418-1ecd4729ac09-kube-api-access-m56mx\") pod \"nova-api-db-create-vk5ks\" (UID: \"9bbdf249-90f9-4685-9418-1ecd4729ac09\") " pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.150939 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bbdf249-90f9-4685-9418-1ecd4729ac09-operator-scripts\") pod \"nova-api-db-create-vk5ks\" (UID: \"9bbdf249-90f9-4685-9418-1ecd4729ac09\") " pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.167633 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ztq89"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.190513 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m56mx\" (UniqueName: \"kubernetes.io/projected/9bbdf249-90f9-4685-9418-1ecd4729ac09-kube-api-access-m56mx\") pod \"nova-api-db-create-vk5ks\" (UID: \"9bbdf249-90f9-4685-9418-1ecd4729ac09\") " pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.230360 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.251069 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d45f2b81-a9f9-47bc-8042-c344bffe4dac-operator-scripts\") pod \"nova-cell1-db-create-ztq89\" (UID: \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\") " pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.251415 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwbsg\" (UniqueName: \"kubernetes.io/projected/50ad724d-2883-415e-97f4-636ee63d864c-kube-api-access-gwbsg\") pod \"nova-cell0-6ac2-account-create-update-l2v8g\" (UID: \"50ad724d-2883-415e-97f4-636ee63d864c\") " pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.251466 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgcf2\" (UniqueName: \"kubernetes.io/projected/d45f2b81-a9f9-47bc-8042-c344bffe4dac-kube-api-access-sgcf2\") pod \"nova-cell1-db-create-ztq89\" (UID: \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\") " pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.251530 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-629cg\" (UniqueName: \"kubernetes.io/projected/75b050ca-b135-4564-be78-fd2a4efefdbe-kube-api-access-629cg\") pod \"nova-cell0-db-create-r28fb\" (UID: \"75b050ca-b135-4564-be78-fd2a4efefdbe\") " pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.251553 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b050ca-b135-4564-be78-fd2a4efefdbe-operator-scripts\") pod \"nova-cell0-db-create-r28fb\" (UID: \"75b050ca-b135-4564-be78-fd2a4efefdbe\") " pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.251585 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56tmh\" (UniqueName: \"kubernetes.io/projected/e8436db7-0c0c-4c64-aea2-2266437d9d44-kube-api-access-56tmh\") pod \"nova-api-5c0f-account-create-update-2gwwc\" (UID: \"e8436db7-0c0c-4c64-aea2-2266437d9d44\") " pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.251641 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50ad724d-2883-415e-97f4-636ee63d864c-operator-scripts\") pod \"nova-cell0-6ac2-account-create-update-l2v8g\" (UID: \"50ad724d-2883-415e-97f4-636ee63d864c\") " pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.251664 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8436db7-0c0c-4c64-aea2-2266437d9d44-operator-scripts\") pod \"nova-api-5c0f-account-create-update-2gwwc\" (UID: \"e8436db7-0c0c-4c64-aea2-2266437d9d44\") " pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.252290 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8436db7-0c0c-4c64-aea2-2266437d9d44-operator-scripts\") pod \"nova-api-5c0f-account-create-update-2gwwc\" (UID: \"e8436db7-0c0c-4c64-aea2-2266437d9d44\") " pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.266907 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56tmh\" (UniqueName: \"kubernetes.io/projected/e8436db7-0c0c-4c64-aea2-2266437d9d44-kube-api-access-56tmh\") pod \"nova-api-5c0f-account-create-update-2gwwc\" (UID: \"e8436db7-0c0c-4c64-aea2-2266437d9d44\") " pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.286186 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.318511 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-a59c-account-create-update-wkkp2"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.319719 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.326113 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.329270 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a59c-account-create-update-wkkp2"] Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.353191 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d45f2b81-a9f9-47bc-8042-c344bffe4dac-operator-scripts\") pod \"nova-cell1-db-create-ztq89\" (UID: \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\") " pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.353239 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwbsg\" (UniqueName: \"kubernetes.io/projected/50ad724d-2883-415e-97f4-636ee63d864c-kube-api-access-gwbsg\") pod \"nova-cell0-6ac2-account-create-update-l2v8g\" (UID: \"50ad724d-2883-415e-97f4-636ee63d864c\") " pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.353294 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgcf2\" (UniqueName: \"kubernetes.io/projected/d45f2b81-a9f9-47bc-8042-c344bffe4dac-kube-api-access-sgcf2\") pod \"nova-cell1-db-create-ztq89\" (UID: \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\") " pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.353330 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-629cg\" (UniqueName: \"kubernetes.io/projected/75b050ca-b135-4564-be78-fd2a4efefdbe-kube-api-access-629cg\") pod \"nova-cell0-db-create-r28fb\" (UID: \"75b050ca-b135-4564-be78-fd2a4efefdbe\") " pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.353353 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b050ca-b135-4564-be78-fd2a4efefdbe-operator-scripts\") pod \"nova-cell0-db-create-r28fb\" (UID: \"75b050ca-b135-4564-be78-fd2a4efefdbe\") " pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.353429 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50ad724d-2883-415e-97f4-636ee63d864c-operator-scripts\") pod \"nova-cell0-6ac2-account-create-update-l2v8g\" (UID: \"50ad724d-2883-415e-97f4-636ee63d864c\") " pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.354148 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50ad724d-2883-415e-97f4-636ee63d864c-operator-scripts\") pod \"nova-cell0-6ac2-account-create-update-l2v8g\" (UID: \"50ad724d-2883-415e-97f4-636ee63d864c\") " pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.354590 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d45f2b81-a9f9-47bc-8042-c344bffe4dac-operator-scripts\") pod \"nova-cell1-db-create-ztq89\" (UID: \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\") " pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.355581 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b050ca-b135-4564-be78-fd2a4efefdbe-operator-scripts\") pod \"nova-cell0-db-create-r28fb\" (UID: \"75b050ca-b135-4564-be78-fd2a4efefdbe\") " pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.376603 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwbsg\" (UniqueName: \"kubernetes.io/projected/50ad724d-2883-415e-97f4-636ee63d864c-kube-api-access-gwbsg\") pod \"nova-cell0-6ac2-account-create-update-l2v8g\" (UID: \"50ad724d-2883-415e-97f4-636ee63d864c\") " pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.380406 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-629cg\" (UniqueName: \"kubernetes.io/projected/75b050ca-b135-4564-be78-fd2a4efefdbe-kube-api-access-629cg\") pod \"nova-cell0-db-create-r28fb\" (UID: \"75b050ca-b135-4564-be78-fd2a4efefdbe\") " pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.394492 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgcf2\" (UniqueName: \"kubernetes.io/projected/d45f2b81-a9f9-47bc-8042-c344bffe4dac-kube-api-access-sgcf2\") pod \"nova-cell1-db-create-ztq89\" (UID: \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\") " pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.455374 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e72addf-edb3-4704-be2e-206825774d97-operator-scripts\") pod \"nova-cell1-a59c-account-create-update-wkkp2\" (UID: \"4e72addf-edb3-4704-be2e-206825774d97\") " pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.455601 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj2xl\" (UniqueName: \"kubernetes.io/projected/4e72addf-edb3-4704-be2e-206825774d97-kube-api-access-dj2xl\") pod \"nova-cell1-a59c-account-create-update-wkkp2\" (UID: \"4e72addf-edb3-4704-be2e-206825774d97\") " pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.462398 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.475556 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.557094 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e72addf-edb3-4704-be2e-206825774d97-operator-scripts\") pod \"nova-cell1-a59c-account-create-update-wkkp2\" (UID: \"4e72addf-edb3-4704-be2e-206825774d97\") " pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.557210 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj2xl\" (UniqueName: \"kubernetes.io/projected/4e72addf-edb3-4704-be2e-206825774d97-kube-api-access-dj2xl\") pod \"nova-cell1-a59c-account-create-update-wkkp2\" (UID: \"4e72addf-edb3-4704-be2e-206825774d97\") " pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.557760 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e72addf-edb3-4704-be2e-206825774d97-operator-scripts\") pod \"nova-cell1-a59c-account-create-update-wkkp2\" (UID: \"4e72addf-edb3-4704-be2e-206825774d97\") " pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.576459 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj2xl\" (UniqueName: \"kubernetes.io/projected/4e72addf-edb3-4704-be2e-206825774d97-kube-api-access-dj2xl\") pod \"nova-cell1-a59c-account-create-update-wkkp2\" (UID: \"4e72addf-edb3-4704-be2e-206825774d97\") " pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.641607 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.654988 4733 generic.go:334] "Generic (PLEG): container finished" podID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerID="262ab0c4956d796d1019c1aab9759043a7cd8f1e971e1c279de42fb81935131d" exitCode=0 Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.655062 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-98f7dbcd8-vdtg8" event={"ID":"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab","Type":"ContainerDied","Data":"262ab0c4956d796d1019c1aab9759043a7cd8f1e971e1c279de42fb81935131d"} Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.657865 4733 generic.go:334] "Generic (PLEG): container finished" podID="527118fc-11ba-4c6b-a15a-415912a5e650" containerID="db6f3a34d675cc354963a482ff62b85e52c39c97b79d0c7f54bfdaf0383617c0" exitCode=143 Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.657967 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"527118fc-11ba-4c6b-a15a-415912a5e650","Type":"ContainerDied","Data":"db6f3a34d675cc354963a482ff62b85e52c39c97b79d0c7f54bfdaf0383617c0"} Dec 04 18:01:09 crc kubenswrapper[4733]: I1204 18:01:09.678157 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:11 crc kubenswrapper[4733]: I1204 18:01:11.332078 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:01:11 crc kubenswrapper[4733]: I1204 18:01:11.370309 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.237578 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.312188 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-run-httpd\") pod \"d6373063-26a1-4f42-9d95-79c54a11f8dc\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.312268 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-sg-core-conf-yaml\") pod \"d6373063-26a1-4f42-9d95-79c54a11f8dc\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.312310 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-scripts\") pod \"d6373063-26a1-4f42-9d95-79c54a11f8dc\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.312333 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-combined-ca-bundle\") pod \"d6373063-26a1-4f42-9d95-79c54a11f8dc\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.312359 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-log-httpd\") pod \"d6373063-26a1-4f42-9d95-79c54a11f8dc\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.312511 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxnlt\" (UniqueName: \"kubernetes.io/projected/d6373063-26a1-4f42-9d95-79c54a11f8dc-kube-api-access-sxnlt\") pod \"d6373063-26a1-4f42-9d95-79c54a11f8dc\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.312546 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-config-data\") pod \"d6373063-26a1-4f42-9d95-79c54a11f8dc\" (UID: \"d6373063-26a1-4f42-9d95-79c54a11f8dc\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.313894 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d6373063-26a1-4f42-9d95-79c54a11f8dc" (UID: "d6373063-26a1-4f42-9d95-79c54a11f8dc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.314249 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d6373063-26a1-4f42-9d95-79c54a11f8dc" (UID: "d6373063-26a1-4f42-9d95-79c54a11f8dc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.317248 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-scripts" (OuterVolumeSpecName: "scripts") pod "d6373063-26a1-4f42-9d95-79c54a11f8dc" (UID: "d6373063-26a1-4f42-9d95-79c54a11f8dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.321493 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6373063-26a1-4f42-9d95-79c54a11f8dc-kube-api-access-sxnlt" (OuterVolumeSpecName: "kube-api-access-sxnlt") pod "d6373063-26a1-4f42-9d95-79c54a11f8dc" (UID: "d6373063-26a1-4f42-9d95-79c54a11f8dc"). InnerVolumeSpecName "kube-api-access-sxnlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.345649 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d6373063-26a1-4f42-9d95-79c54a11f8dc" (UID: "d6373063-26a1-4f42-9d95-79c54a11f8dc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.384174 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.413758 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-combined-ca-bundle\") pod \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414038 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-config\") pod \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414170 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-httpd-config\") pod \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414189 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxzdf\" (UniqueName: \"kubernetes.io/projected/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-kube-api-access-kxzdf\") pod \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414211 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-ovndb-tls-certs\") pod \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\" (UID: \"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414635 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxnlt\" (UniqueName: \"kubernetes.io/projected/d6373063-26a1-4f42-9d95-79c54a11f8dc-kube-api-access-sxnlt\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414645 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414654 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414662 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.414670 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6373063-26a1-4f42-9d95-79c54a11f8dc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.426159 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" (UID: "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.426293 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-kube-api-access-kxzdf" (OuterVolumeSpecName: "kube-api-access-kxzdf") pod "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" (UID: "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab"). InnerVolumeSpecName "kube-api-access-kxzdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.439772 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6373063-26a1-4f42-9d95-79c54a11f8dc" (UID: "d6373063-26a1-4f42-9d95-79c54a11f8dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.473983 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-config-data" (OuterVolumeSpecName: "config-data") pod "d6373063-26a1-4f42-9d95-79c54a11f8dc" (UID: "d6373063-26a1-4f42-9d95-79c54a11f8dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.478818 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" (UID: "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.479929 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-config" (OuterVolumeSpecName: "config") pod "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" (UID: "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.499382 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" (UID: "4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.516269 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.516523 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.516603 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxzdf\" (UniqueName: \"kubernetes.io/projected/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-kube-api-access-kxzdf\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.516658 4733 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.516730 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6373063-26a1-4f42-9d95-79c54a11f8dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.516827 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.516882 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.694664 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-98f7dbcd8-vdtg8" event={"ID":"4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab","Type":"ContainerDied","Data":"a2e6fac54a9ac5724cc35da16d710ef5e2a58c20e7b09bd709b6a6a77ce42cef"} Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.694712 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-98f7dbcd8-vdtg8" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.694726 4733 scope.go:117] "RemoveContainer" containerID="5ae8831a27bc1dd563aa1309ef400056f32ce6d82dc70a522d2f89456bde35d9" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.701009 4733 generic.go:334] "Generic (PLEG): container finished" podID="527118fc-11ba-4c6b-a15a-415912a5e650" containerID="ec9864790923d6c96b26352e7b30f7d1edab920785e6bd453aae8436fb4d67ab" exitCode=0 Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.701101 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"527118fc-11ba-4c6b-a15a-415912a5e650","Type":"ContainerDied","Data":"ec9864790923d6c96b26352e7b30f7d1edab920785e6bd453aae8436fb4d67ab"} Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.704141 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6373063-26a1-4f42-9d95-79c54a11f8dc","Type":"ContainerDied","Data":"089644e6f55e6c998efe4294a200d397d4ca51cd608492a5c720a1e760ae6d2a"} Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.704220 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.721245 4733 scope.go:117] "RemoveContainer" containerID="262ab0c4956d796d1019c1aab9759043a7cd8f1e971e1c279de42fb81935131d" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.883239 4733 scope.go:117] "RemoveContainer" containerID="b8dd0239fe51987da1c26aa092d25f2f708220112e9e888bc7a2dc790b0ac8bb" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.888577 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.905122 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-r28fb"] Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.909063 4733 scope.go:117] "RemoveContainer" containerID="b8d50c7b9a90929781ea8fa565d16c51cf9b07761341fe1fcc13b62cd52ebfa4" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.918344 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5c0f-account-create-update-2gwwc"] Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.922088 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-logs\") pod \"527118fc-11ba-4c6b-a15a-415912a5e650\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.922181 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-httpd-run\") pod \"527118fc-11ba-4c6b-a15a-415912a5e650\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.922228 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-combined-ca-bundle\") pod \"527118fc-11ba-4c6b-a15a-415912a5e650\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.922274 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"527118fc-11ba-4c6b-a15a-415912a5e650\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.922331 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-public-tls-certs\") pod \"527118fc-11ba-4c6b-a15a-415912a5e650\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.922354 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-scripts\") pod \"527118fc-11ba-4c6b-a15a-415912a5e650\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.922391 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-config-data\") pod \"527118fc-11ba-4c6b-a15a-415912a5e650\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.922413 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf8hb\" (UniqueName: \"kubernetes.io/projected/527118fc-11ba-4c6b-a15a-415912a5e650-kube-api-access-cf8hb\") pod \"527118fc-11ba-4c6b-a15a-415912a5e650\" (UID: \"527118fc-11ba-4c6b-a15a-415912a5e650\") " Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.924042 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "527118fc-11ba-4c6b-a15a-415912a5e650" (UID: "527118fc-11ba-4c6b-a15a-415912a5e650"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.924534 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-logs" (OuterVolumeSpecName: "logs") pod "527118fc-11ba-4c6b-a15a-415912a5e650" (UID: "527118fc-11ba-4c6b-a15a-415912a5e650"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.929395 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "527118fc-11ba-4c6b-a15a-415912a5e650" (UID: "527118fc-11ba-4c6b-a15a-415912a5e650"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.929642 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527118fc-11ba-4c6b-a15a-415912a5e650-kube-api-access-cf8hb" (OuterVolumeSpecName: "kube-api-access-cf8hb") pod "527118fc-11ba-4c6b-a15a-415912a5e650" (UID: "527118fc-11ba-4c6b-a15a-415912a5e650"). InnerVolumeSpecName "kube-api-access-cf8hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.930244 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ztq89"] Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.932477 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-scripts" (OuterVolumeSpecName: "scripts") pod "527118fc-11ba-4c6b-a15a-415912a5e650" (UID: "527118fc-11ba-4c6b-a15a-415912a5e650"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.945577 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-98f7dbcd8-vdtg8"] Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.956314 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-98f7dbcd8-vdtg8"] Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.974993 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.980257 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vk5ks"] Dec 04 18:01:12 crc kubenswrapper[4733]: I1204 18:01:12.994089 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6ac2-account-create-update-l2v8g"] Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.012666 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "527118fc-11ba-4c6b-a15a-415912a5e650" (UID: "527118fc-11ba-4c6b-a15a-415912a5e650"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.031368 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.031400 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.031411 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf8hb\" (UniqueName: \"kubernetes.io/projected/527118fc-11ba-4c6b-a15a-415912a5e650-kube-api-access-cf8hb\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.031424 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.031432 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527118fc-11ba-4c6b-a15a-415912a5e650-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.031455 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.063128 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.073856 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.085036 4733 scope.go:117] "RemoveContainer" containerID="3d2f01ff68acc6abde2f03f1dfbcf4dbaf75f00f013effb036662df16bbc7102" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091036 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:13 crc kubenswrapper[4733]: E1204 18:01:13.091474 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="ceilometer-notification-agent" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091494 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="ceilometer-notification-agent" Dec 04 18:01:13 crc kubenswrapper[4733]: E1204 18:01:13.091506 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="proxy-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091513 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="proxy-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: E1204 18:01:13.091522 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerName="neutron-api" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091528 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerName="neutron-api" Dec 04 18:01:13 crc kubenswrapper[4733]: E1204 18:01:13.091544 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="ceilometer-central-agent" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091550 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="ceilometer-central-agent" Dec 04 18:01:13 crc kubenswrapper[4733]: E1204 18:01:13.091570 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" containerName="glance-log" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091575 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" containerName="glance-log" Dec 04 18:01:13 crc kubenswrapper[4733]: E1204 18:01:13.091581 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerName="neutron-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091587 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerName="neutron-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: E1204 18:01:13.091604 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" containerName="glance-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091609 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" containerName="glance-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: E1204 18:01:13.091619 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="sg-core" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091625 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="sg-core" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091743 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "527118fc-11ba-4c6b-a15a-415912a5e650" (UID: "527118fc-11ba-4c6b-a15a-415912a5e650"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091774 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerName="neutron-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091785 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="proxy-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091810 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" containerName="glance-log" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091819 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" containerName="neutron-api" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091835 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="sg-core" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091842 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="ceilometer-central-agent" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091856 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" containerName="ceilometer-notification-agent" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.091866 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" containerName="glance-httpd" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.093688 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.099055 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.100523 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.113977 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-config-data" (OuterVolumeSpecName: "config-data") pod "527118fc-11ba-4c6b-a15a-415912a5e650" (UID: "527118fc-11ba-4c6b-a15a-415912a5e650"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.127364 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a59c-account-create-update-wkkp2"] Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.131806 4733 scope.go:117] "RemoveContainer" containerID="a9a9a9cdb9dcefd329ac3cb110ec4b2683940979f87de4e19d3577cc0127994b" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.132768 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-config-data\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.132832 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5n88\" (UniqueName: \"kubernetes.io/projected/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-kube-api-access-r5n88\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.132851 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.132883 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-log-httpd\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.132900 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-run-httpd\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.132958 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-scripts\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.133030 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.133080 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.133090 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.133100 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527118fc-11ba-4c6b-a15a-415912a5e650-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.136780 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.234310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.234351 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-config-data\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.234389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5n88\" (UniqueName: \"kubernetes.io/projected/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-kube-api-access-r5n88\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.234406 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.234436 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-log-httpd\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.234453 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-run-httpd\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.234479 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-scripts\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.235360 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-log-httpd\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.235369 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-run-httpd\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.240244 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-scripts\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.241305 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.241514 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.248316 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-config-data\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.256267 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5n88\" (UniqueName: \"kubernetes.io/projected/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-kube-api-access-r5n88\") pod \"ceilometer-0\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.358572 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.358862 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerName="glance-log" containerID="cri-o://1241cadb321985565a1841b5836435da7c72c4337f283fd97ea244d6cfe421d2" gracePeriod=30 Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.358937 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerName="glance-httpd" containerID="cri-o://5c0ddd09e462eda1e481a8b62bdcbb3fee36cd52d17f119c18c2049111ec6377" gracePeriod=30 Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.419423 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.420240 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.747590 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerID="1241cadb321985565a1841b5836435da7c72c4337f283fd97ea244d6cfe421d2" exitCode=143 Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.747659 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4f82a6d-dcad-432e-a365-38df92ff4e6e","Type":"ContainerDied","Data":"1241cadb321985565a1841b5836435da7c72c4337f283fd97ea244d6cfe421d2"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.755265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" event={"ID":"e8436db7-0c0c-4c64-aea2-2266437d9d44","Type":"ContainerStarted","Data":"011962e29ba8da2577d347e04ff5bb5770e5fc4174508d0f00b388328c73e4f8"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.755316 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" event={"ID":"e8436db7-0c0c-4c64-aea2-2266437d9d44","Type":"ContainerStarted","Data":"a01173a5be3072310766d58c74537c8acfdb0826b5923981a586036da50bc58e"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.760992 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ztq89" event={"ID":"d45f2b81-a9f9-47bc-8042-c344bffe4dac","Type":"ContainerStarted","Data":"279356342a7696e8f66f5a4be3dcb02380255c870e5fc62c09803611ee661684"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.761028 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ztq89" event={"ID":"d45f2b81-a9f9-47bc-8042-c344bffe4dac","Type":"ContainerStarted","Data":"7b29578c5fe06ea93f5e90f143f6292ff386f357aee2ad1b923dc424929cfded"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.765758 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" event={"ID":"50ad724d-2883-415e-97f4-636ee63d864c","Type":"ContainerStarted","Data":"84b5297fb15dad792d9c940a5d4f0d67d6d13f94bd2b0ff3d1ae0a60960f7798"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.765807 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" event={"ID":"50ad724d-2883-415e-97f4-636ee63d864c","Type":"ContainerStarted","Data":"5c60e54eca03f3d9f6e420f341e35154279c7a23dd9fb2d4acbfed13be8d2d25"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.785724 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"7b19ede9-4e23-416d-bc7a-3674ed551c21","Type":"ContainerStarted","Data":"ebae6d3dccdf5b42c07640f9dd249ddf0193a7662a64ce8f92ecd4ca8021c824"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.786463 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-ztq89" podStartSLOduration=4.786444952 podStartE2EDuration="4.786444952s" podCreationTimestamp="2025-12-04 18:01:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:13.785338233 +0000 UTC m=+1335.740699279" watchObservedRunningTime="2025-12-04 18:01:13.786444952 +0000 UTC m=+1335.741806008" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.788524 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" podStartSLOduration=5.788513996 podStartE2EDuration="5.788513996s" podCreationTimestamp="2025-12-04 18:01:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:13.771323714 +0000 UTC m=+1335.726684760" watchObservedRunningTime="2025-12-04 18:01:13.788513996 +0000 UTC m=+1335.743875052" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.801819 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"527118fc-11ba-4c6b-a15a-415912a5e650","Type":"ContainerDied","Data":"5ef2507be5bdc6228b6a16c06d1aaf7a8d712c4efea9ff32778800be6bbef7c3"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.801964 4733 scope.go:117] "RemoveContainer" containerID="ec9864790923d6c96b26352e7b30f7d1edab920785e6bd453aae8436fb4d67ab" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.803883 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.808216 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vk5ks" event={"ID":"9bbdf249-90f9-4685-9418-1ecd4729ac09","Type":"ContainerStarted","Data":"e20b5cfb354e7b5111ecacb69236c810870c0447a1bda743bdcf9810c1519d94"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.808252 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vk5ks" event={"ID":"9bbdf249-90f9-4685-9418-1ecd4729ac09","Type":"ContainerStarted","Data":"ea527dc088d8a21f729429548a65fef15f644b60b24f462a468184f6125fac12"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.811913 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" podStartSLOduration=4.811899381 podStartE2EDuration="4.811899381s" podCreationTimestamp="2025-12-04 18:01:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:13.802053163 +0000 UTC m=+1335.757414209" watchObservedRunningTime="2025-12-04 18:01:13.811899381 +0000 UTC m=+1335.767260427" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.814615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" event={"ID":"4e72addf-edb3-4704-be2e-206825774d97","Type":"ContainerStarted","Data":"a3eef607b3353284902508f0ced38edd93dbee9370ecc514d4db535fa52597df"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.814682 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" event={"ID":"4e72addf-edb3-4704-be2e-206825774d97","Type":"ContainerStarted","Data":"db96e4afc49c67c49822536e5acd3e352c3e078d183f3e8faaa317b7fd36d6a2"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.837980 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.745062483 podStartE2EDuration="14.837921956s" podCreationTimestamp="2025-12-04 18:00:59 +0000 UTC" firstStartedPulling="2025-12-04 18:01:00.022607568 +0000 UTC m=+1321.977968604" lastFinishedPulling="2025-12-04 18:01:12.115467031 +0000 UTC m=+1334.070828077" observedRunningTime="2025-12-04 18:01:13.823814384 +0000 UTC m=+1335.779175470" watchObservedRunningTime="2025-12-04 18:01:13.837921956 +0000 UTC m=+1335.793283022" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.838178 4733 scope.go:117] "RemoveContainer" containerID="db6f3a34d675cc354963a482ff62b85e52c39c97b79d0c7f54bfdaf0383617c0" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.841087 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r28fb" event={"ID":"75b050ca-b135-4564-be78-fd2a4efefdbe","Type":"ContainerStarted","Data":"da28237687950c4a080dd5e279c384c4f78082957a7be8c3e467ac127824eb2b"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.841134 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r28fb" event={"ID":"75b050ca-b135-4564-be78-fd2a4efefdbe","Type":"ContainerStarted","Data":"c33d06e0355597b1c9b8d44da9f75a68744763436c43f8748bfe7c88f8b0167b"} Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.867408 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" podStartSLOduration=4.86738621 podStartE2EDuration="4.86738621s" podCreationTimestamp="2025-12-04 18:01:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:13.851613285 +0000 UTC m=+1335.806974331" watchObservedRunningTime="2025-12-04 18:01:13.86738621 +0000 UTC m=+1335.822747256" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.877276 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-vk5ks" podStartSLOduration=5.877256 podStartE2EDuration="5.877256s" podCreationTimestamp="2025-12-04 18:01:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:13.868123989 +0000 UTC m=+1335.823485055" watchObservedRunningTime="2025-12-04 18:01:13.877256 +0000 UTC m=+1335.832617046" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.895837 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.898296 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-r28fb" podStartSLOduration=5.898272112 podStartE2EDuration="5.898272112s" podCreationTimestamp="2025-12-04 18:01:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:13.88793911 +0000 UTC m=+1335.843300146" watchObservedRunningTime="2025-12-04 18:01:13.898272112 +0000 UTC m=+1335.853633158" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.907485 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.912437 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:13 crc kubenswrapper[4733]: I1204 18:01:13.933717 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.005184 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.032082 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.037664 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.039846 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.040448 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.044199 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.157372 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.157449 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.157528 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.157557 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hhhk\" (UniqueName: \"kubernetes.io/projected/2c8b0056-e347-4f24-aeea-20680c33bf06-kube-api-access-7hhhk\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.157582 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.157629 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.157756 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.157778 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-logs\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259174 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259218 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-logs\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259267 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259370 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259398 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hhhk\" (UniqueName: \"kubernetes.io/projected/2c8b0056-e347-4f24-aeea-20680c33bf06-kube-api-access-7hhhk\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259420 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259457 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259856 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-logs\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.259888 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.260403 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.274650 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.275185 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.275672 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.277542 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.280142 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hhhk\" (UniqueName: \"kubernetes.io/projected/2c8b0056-e347-4f24-aeea-20680c33bf06-kube-api-access-7hhhk\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.296295 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.350916 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab" path="/var/lib/kubelet/pods/4bc84a76-c1da-4cb4-ba16-a3aeac81c1ab/volumes" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.351546 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527118fc-11ba-4c6b-a15a-415912a5e650" path="/var/lib/kubelet/pods/527118fc-11ba-4c6b-a15a-415912a5e650/volumes" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.352130 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6373063-26a1-4f42-9d95-79c54a11f8dc" path="/var/lib/kubelet/pods/d6373063-26a1-4f42-9d95-79c54a11f8dc/volumes" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.489085 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.852414 4733 generic.go:334] "Generic (PLEG): container finished" podID="75b050ca-b135-4564-be78-fd2a4efefdbe" containerID="da28237687950c4a080dd5e279c384c4f78082957a7be8c3e467ac127824eb2b" exitCode=0 Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.852512 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r28fb" event={"ID":"75b050ca-b135-4564-be78-fd2a4efefdbe","Type":"ContainerDied","Data":"da28237687950c4a080dd5e279c384c4f78082957a7be8c3e467ac127824eb2b"} Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.853838 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerStarted","Data":"57eb48e9a4d76efb1c5b1071fcebfc100ff53894ee075a47ebb4971ed948727c"} Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.855370 4733 generic.go:334] "Generic (PLEG): container finished" podID="50ad724d-2883-415e-97f4-636ee63d864c" containerID="84b5297fb15dad792d9c940a5d4f0d67d6d13f94bd2b0ff3d1ae0a60960f7798" exitCode=0 Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.855431 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" event={"ID":"50ad724d-2883-415e-97f4-636ee63d864c","Type":"ContainerDied","Data":"84b5297fb15dad792d9c940a5d4f0d67d6d13f94bd2b0ff3d1ae0a60960f7798"} Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.858526 4733 generic.go:334] "Generic (PLEG): container finished" podID="9bbdf249-90f9-4685-9418-1ecd4729ac09" containerID="e20b5cfb354e7b5111ecacb69236c810870c0447a1bda743bdcf9810c1519d94" exitCode=0 Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.858625 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vk5ks" event={"ID":"9bbdf249-90f9-4685-9418-1ecd4729ac09","Type":"ContainerDied","Data":"e20b5cfb354e7b5111ecacb69236c810870c0447a1bda743bdcf9810c1519d94"} Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.860305 4733 generic.go:334] "Generic (PLEG): container finished" podID="4e72addf-edb3-4704-be2e-206825774d97" containerID="a3eef607b3353284902508f0ced38edd93dbee9370ecc514d4db535fa52597df" exitCode=0 Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.860372 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" event={"ID":"4e72addf-edb3-4704-be2e-206825774d97","Type":"ContainerDied","Data":"a3eef607b3353284902508f0ced38edd93dbee9370ecc514d4db535fa52597df"} Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.861598 4733 generic.go:334] "Generic (PLEG): container finished" podID="e8436db7-0c0c-4c64-aea2-2266437d9d44" containerID="011962e29ba8da2577d347e04ff5bb5770e5fc4174508d0f00b388328c73e4f8" exitCode=0 Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.861662 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" event={"ID":"e8436db7-0c0c-4c64-aea2-2266437d9d44","Type":"ContainerDied","Data":"011962e29ba8da2577d347e04ff5bb5770e5fc4174508d0f00b388328c73e4f8"} Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.866199 4733 generic.go:334] "Generic (PLEG): container finished" podID="d45f2b81-a9f9-47bc-8042-c344bffe4dac" containerID="279356342a7696e8f66f5a4be3dcb02380255c870e5fc62c09803611ee661684" exitCode=0 Dec 04 18:01:14 crc kubenswrapper[4733]: I1204 18:01:14.866290 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ztq89" event={"ID":"d45f2b81-a9f9-47bc-8042-c344bffe4dac","Type":"ContainerDied","Data":"279356342a7696e8f66f5a4be3dcb02380255c870e5fc62c09803611ee661684"} Dec 04 18:01:15 crc kubenswrapper[4733]: I1204 18:01:15.084515 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:01:15 crc kubenswrapper[4733]: I1204 18:01:15.884864 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c8b0056-e347-4f24-aeea-20680c33bf06","Type":"ContainerStarted","Data":"63006d63457f2e071c6d109c86778556fbf386e34e8d0116bd25dc63bd13509f"} Dec 04 18:01:15 crc kubenswrapper[4733]: I1204 18:01:15.885332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c8b0056-e347-4f24-aeea-20680c33bf06","Type":"ContainerStarted","Data":"00657a27d84c4ba431e0def2a7ed219f0b276dd285ad90f3864cea5721e2fe7d"} Dec 04 18:01:15 crc kubenswrapper[4733]: I1204 18:01:15.888069 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerStarted","Data":"c095546f6417c9108d455c9a1598286cc75cac04c57dee19b93698265f2c6976"} Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.380600 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.516301 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj2xl\" (UniqueName: \"kubernetes.io/projected/4e72addf-edb3-4704-be2e-206825774d97-kube-api-access-dj2xl\") pod \"4e72addf-edb3-4704-be2e-206825774d97\" (UID: \"4e72addf-edb3-4704-be2e-206825774d97\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.516735 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e72addf-edb3-4704-be2e-206825774d97-operator-scripts\") pod \"4e72addf-edb3-4704-be2e-206825774d97\" (UID: \"4e72addf-edb3-4704-be2e-206825774d97\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.517885 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e72addf-edb3-4704-be2e-206825774d97-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4e72addf-edb3-4704-be2e-206825774d97" (UID: "4e72addf-edb3-4704-be2e-206825774d97"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.533136 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e72addf-edb3-4704-be2e-206825774d97-kube-api-access-dj2xl" (OuterVolumeSpecName: "kube-api-access-dj2xl") pod "4e72addf-edb3-4704-be2e-206825774d97" (UID: "4e72addf-edb3-4704-be2e-206825774d97"). InnerVolumeSpecName "kube-api-access-dj2xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.618975 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj2xl\" (UniqueName: \"kubernetes.io/projected/4e72addf-edb3-4704-be2e-206825774d97-kube-api-access-dj2xl\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.619008 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e72addf-edb3-4704-be2e-206825774d97-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.621529 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.627240 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.634006 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.691456 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.693605 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.720210 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56tmh\" (UniqueName: \"kubernetes.io/projected/e8436db7-0c0c-4c64-aea2-2266437d9d44-kube-api-access-56tmh\") pod \"e8436db7-0c0c-4c64-aea2-2266437d9d44\" (UID: \"e8436db7-0c0c-4c64-aea2-2266437d9d44\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.720369 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgcf2\" (UniqueName: \"kubernetes.io/projected/d45f2b81-a9f9-47bc-8042-c344bffe4dac-kube-api-access-sgcf2\") pod \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\" (UID: \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.720429 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b050ca-b135-4564-be78-fd2a4efefdbe-operator-scripts\") pod \"75b050ca-b135-4564-be78-fd2a4efefdbe\" (UID: \"75b050ca-b135-4564-be78-fd2a4efefdbe\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.720592 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-629cg\" (UniqueName: \"kubernetes.io/projected/75b050ca-b135-4564-be78-fd2a4efefdbe-kube-api-access-629cg\") pod \"75b050ca-b135-4564-be78-fd2a4efefdbe\" (UID: \"75b050ca-b135-4564-be78-fd2a4efefdbe\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.720734 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8436db7-0c0c-4c64-aea2-2266437d9d44-operator-scripts\") pod \"e8436db7-0c0c-4c64-aea2-2266437d9d44\" (UID: \"e8436db7-0c0c-4c64-aea2-2266437d9d44\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.720787 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d45f2b81-a9f9-47bc-8042-c344bffe4dac-operator-scripts\") pod \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\" (UID: \"d45f2b81-a9f9-47bc-8042-c344bffe4dac\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.721442 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8436db7-0c0c-4c64-aea2-2266437d9d44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8436db7-0c0c-4c64-aea2-2266437d9d44" (UID: "e8436db7-0c0c-4c64-aea2-2266437d9d44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.721520 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75b050ca-b135-4564-be78-fd2a4efefdbe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "75b050ca-b135-4564-be78-fd2a4efefdbe" (UID: "75b050ca-b135-4564-be78-fd2a4efefdbe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.721882 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d45f2b81-a9f9-47bc-8042-c344bffe4dac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d45f2b81-a9f9-47bc-8042-c344bffe4dac" (UID: "d45f2b81-a9f9-47bc-8042-c344bffe4dac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.723895 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d45f2b81-a9f9-47bc-8042-c344bffe4dac-kube-api-access-sgcf2" (OuterVolumeSpecName: "kube-api-access-sgcf2") pod "d45f2b81-a9f9-47bc-8042-c344bffe4dac" (UID: "d45f2b81-a9f9-47bc-8042-c344bffe4dac"). InnerVolumeSpecName "kube-api-access-sgcf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.727189 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8436db7-0c0c-4c64-aea2-2266437d9d44-kube-api-access-56tmh" (OuterVolumeSpecName: "kube-api-access-56tmh") pod "e8436db7-0c0c-4c64-aea2-2266437d9d44" (UID: "e8436db7-0c0c-4c64-aea2-2266437d9d44"). InnerVolumeSpecName "kube-api-access-56tmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.743646 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75b050ca-b135-4564-be78-fd2a4efefdbe-kube-api-access-629cg" (OuterVolumeSpecName: "kube-api-access-629cg") pod "75b050ca-b135-4564-be78-fd2a4efefdbe" (UID: "75b050ca-b135-4564-be78-fd2a4efefdbe"). InnerVolumeSpecName "kube-api-access-629cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.825189 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m56mx\" (UniqueName: \"kubernetes.io/projected/9bbdf249-90f9-4685-9418-1ecd4729ac09-kube-api-access-m56mx\") pod \"9bbdf249-90f9-4685-9418-1ecd4729ac09\" (UID: \"9bbdf249-90f9-4685-9418-1ecd4729ac09\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.825558 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50ad724d-2883-415e-97f4-636ee63d864c-operator-scripts\") pod \"50ad724d-2883-415e-97f4-636ee63d864c\" (UID: \"50ad724d-2883-415e-97f4-636ee63d864c\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.825622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bbdf249-90f9-4685-9418-1ecd4729ac09-operator-scripts\") pod \"9bbdf249-90f9-4685-9418-1ecd4729ac09\" (UID: \"9bbdf249-90f9-4685-9418-1ecd4729ac09\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.825694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwbsg\" (UniqueName: \"kubernetes.io/projected/50ad724d-2883-415e-97f4-636ee63d864c-kube-api-access-gwbsg\") pod \"50ad724d-2883-415e-97f4-636ee63d864c\" (UID: \"50ad724d-2883-415e-97f4-636ee63d864c\") " Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826039 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50ad724d-2883-415e-97f4-636ee63d864c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "50ad724d-2883-415e-97f4-636ee63d864c" (UID: "50ad724d-2883-415e-97f4-636ee63d864c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826461 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgcf2\" (UniqueName: \"kubernetes.io/projected/d45f2b81-a9f9-47bc-8042-c344bffe4dac-kube-api-access-sgcf2\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826491 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b050ca-b135-4564-be78-fd2a4efefdbe-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826501 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-629cg\" (UniqueName: \"kubernetes.io/projected/75b050ca-b135-4564-be78-fd2a4efefdbe-kube-api-access-629cg\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826510 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50ad724d-2883-415e-97f4-636ee63d864c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826518 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8436db7-0c0c-4c64-aea2-2266437d9d44-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826527 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d45f2b81-a9f9-47bc-8042-c344bffe4dac-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826535 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56tmh\" (UniqueName: \"kubernetes.io/projected/e8436db7-0c0c-4c64-aea2-2266437d9d44-kube-api-access-56tmh\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.826833 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bbdf249-90f9-4685-9418-1ecd4729ac09-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9bbdf249-90f9-4685-9418-1ecd4729ac09" (UID: "9bbdf249-90f9-4685-9418-1ecd4729ac09"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.828955 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bbdf249-90f9-4685-9418-1ecd4729ac09-kube-api-access-m56mx" (OuterVolumeSpecName: "kube-api-access-m56mx") pod "9bbdf249-90f9-4685-9418-1ecd4729ac09" (UID: "9bbdf249-90f9-4685-9418-1ecd4729ac09"). InnerVolumeSpecName "kube-api-access-m56mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.829291 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50ad724d-2883-415e-97f4-636ee63d864c-kube-api-access-gwbsg" (OuterVolumeSpecName: "kube-api-access-gwbsg") pod "50ad724d-2883-415e-97f4-636ee63d864c" (UID: "50ad724d-2883-415e-97f4-636ee63d864c"). InnerVolumeSpecName "kube-api-access-gwbsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.946934 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bbdf249-90f9-4685-9418-1ecd4729ac09-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.948532 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwbsg\" (UniqueName: \"kubernetes.io/projected/50ad724d-2883-415e-97f4-636ee63d864c-kube-api-access-gwbsg\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.948575 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m56mx\" (UniqueName: \"kubernetes.io/projected/9bbdf249-90f9-4685-9418-1ecd4729ac09-kube-api-access-m56mx\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.968132 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" event={"ID":"e8436db7-0c0c-4c64-aea2-2266437d9d44","Type":"ContainerDied","Data":"a01173a5be3072310766d58c74537c8acfdb0826b5923981a586036da50bc58e"} Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.968333 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a01173a5be3072310766d58c74537c8acfdb0826b5923981a586036da50bc58e" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.968494 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5c0f-account-create-update-2gwwc" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.989212 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ztq89" Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.989206 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ztq89" event={"ID":"d45f2b81-a9f9-47bc-8042-c344bffe4dac","Type":"ContainerDied","Data":"7b29578c5fe06ea93f5e90f143f6292ff386f357aee2ad1b923dc424929cfded"} Dec 04 18:01:16 crc kubenswrapper[4733]: I1204 18:01:16.989937 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b29578c5fe06ea93f5e90f143f6292ff386f357aee2ad1b923dc424929cfded" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.001312 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c8b0056-e347-4f24-aeea-20680c33bf06","Type":"ContainerStarted","Data":"cbcf529109c0c4af1bdf871c346e94fb3c2778251cd7b3b1d1b5864a31229cb3"} Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.018683 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerStarted","Data":"6105deacb60d3518652148ca57b67964adf980319e92bdaaa08f7ca634f506dc"} Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.050955 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vk5ks" event={"ID":"9bbdf249-90f9-4685-9418-1ecd4729ac09","Type":"ContainerDied","Data":"ea527dc088d8a21f729429548a65fef15f644b60b24f462a468184f6125fac12"} Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.051084 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea527dc088d8a21f729429548a65fef15f644b60b24f462a468184f6125fac12" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.051205 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vk5ks" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.076449 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.076431258 podStartE2EDuration="4.076431258s" podCreationTimestamp="2025-12-04 18:01:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:17.051382459 +0000 UTC m=+1339.006743505" watchObservedRunningTime="2025-12-04 18:01:17.076431258 +0000 UTC m=+1339.031792304" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.076844 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" event={"ID":"4e72addf-edb3-4704-be2e-206825774d97","Type":"ContainerDied","Data":"db96e4afc49c67c49822536e5acd3e352c3e078d183f3e8faaa317b7fd36d6a2"} Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.076917 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db96e4afc49c67c49822536e5acd3e352c3e078d183f3e8faaa317b7fd36d6a2" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.076990 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a59c-account-create-update-wkkp2" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.081721 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerID="5c0ddd09e462eda1e481a8b62bdcbb3fee36cd52d17f119c18c2049111ec6377" exitCode=0 Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.081865 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4f82a6d-dcad-432e-a365-38df92ff4e6e","Type":"ContainerDied","Data":"5c0ddd09e462eda1e481a8b62bdcbb3fee36cd52d17f119c18c2049111ec6377"} Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.086003 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r28fb" event={"ID":"75b050ca-b135-4564-be78-fd2a4efefdbe","Type":"ContainerDied","Data":"c33d06e0355597b1c9b8d44da9f75a68744763436c43f8748bfe7c88f8b0167b"} Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.086096 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c33d06e0355597b1c9b8d44da9f75a68744763436c43f8748bfe7c88f8b0167b" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.086207 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r28fb" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.096347 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" event={"ID":"50ad724d-2883-415e-97f4-636ee63d864c","Type":"ContainerDied","Data":"5c60e54eca03f3d9f6e420f341e35154279c7a23dd9fb2d4acbfed13be8d2d25"} Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.096457 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c60e54eca03f3d9f6e420f341e35154279c7a23dd9fb2d4acbfed13be8d2d25" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.096574 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6ac2-account-create-update-l2v8g" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.225071 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.359975 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-scripts\") pod \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.360267 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-combined-ca-bundle\") pod \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.360360 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.360382 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-internal-tls-certs\") pod \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.360411 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-logs\") pod \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.360425 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-config-data\") pod \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.360450 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpfbg\" (UniqueName: \"kubernetes.io/projected/f4f82a6d-dcad-432e-a365-38df92ff4e6e-kube-api-access-dpfbg\") pod \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.360490 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-httpd-run\") pod \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\" (UID: \"f4f82a6d-dcad-432e-a365-38df92ff4e6e\") " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.361121 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f4f82a6d-dcad-432e-a365-38df92ff4e6e" (UID: "f4f82a6d-dcad-432e-a365-38df92ff4e6e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.364118 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-logs" (OuterVolumeSpecName: "logs") pod "f4f82a6d-dcad-432e-a365-38df92ff4e6e" (UID: "f4f82a6d-dcad-432e-a365-38df92ff4e6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.367600 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-scripts" (OuterVolumeSpecName: "scripts") pod "f4f82a6d-dcad-432e-a365-38df92ff4e6e" (UID: "f4f82a6d-dcad-432e-a365-38df92ff4e6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.371520 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "f4f82a6d-dcad-432e-a365-38df92ff4e6e" (UID: "f4f82a6d-dcad-432e-a365-38df92ff4e6e"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.371687 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4f82a6d-dcad-432e-a365-38df92ff4e6e-kube-api-access-dpfbg" (OuterVolumeSpecName: "kube-api-access-dpfbg") pod "f4f82a6d-dcad-432e-a365-38df92ff4e6e" (UID: "f4f82a6d-dcad-432e-a365-38df92ff4e6e"). InnerVolumeSpecName "kube-api-access-dpfbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.397888 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4f82a6d-dcad-432e-a365-38df92ff4e6e" (UID: "f4f82a6d-dcad-432e-a365-38df92ff4e6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.416325 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-config-data" (OuterVolumeSpecName: "config-data") pod "f4f82a6d-dcad-432e-a365-38df92ff4e6e" (UID: "f4f82a6d-dcad-432e-a365-38df92ff4e6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.444628 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f4f82a6d-dcad-432e-a365-38df92ff4e6e" (UID: "f4f82a6d-dcad-432e-a365-38df92ff4e6e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.462473 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.462684 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.462756 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.462810 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.462822 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.462835 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4f82a6d-dcad-432e-a365-38df92ff4e6e-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.462846 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f82a6d-dcad-432e-a365-38df92ff4e6e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.462856 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpfbg\" (UniqueName: \"kubernetes.io/projected/f4f82a6d-dcad-432e-a365-38df92ff4e6e-kube-api-access-dpfbg\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.485897 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 04 18:01:17 crc kubenswrapper[4733]: I1204 18:01:17.564140 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.107531 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerStarted","Data":"2edb7919ef80734ff54d234ee326e8e503ecf46f90a48fb930a58d621749bfa2"} Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.111229 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4f82a6d-dcad-432e-a365-38df92ff4e6e","Type":"ContainerDied","Data":"84156c3fe19ccf80726b8ee78e8b0c7f64e95fe849240198a520dced13ef8220"} Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.111389 4733 scope.go:117] "RemoveContainer" containerID="5c0ddd09e462eda1e481a8b62bdcbb3fee36cd52d17f119c18c2049111ec6377" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.111724 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.137076 4733 scope.go:117] "RemoveContainer" containerID="1241cadb321985565a1841b5836435da7c72c4337f283fd97ea244d6cfe421d2" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.157481 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.157538 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.215269 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:01:18 crc kubenswrapper[4733]: E1204 18:01:18.215891 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ad724d-2883-415e-97f4-636ee63d864c" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.215919 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ad724d-2883-415e-97f4-636ee63d864c" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: E1204 18:01:18.215937 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerName="glance-log" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.215951 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerName="glance-log" Dec 04 18:01:18 crc kubenswrapper[4733]: E1204 18:01:18.215981 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b050ca-b135-4564-be78-fd2a4efefdbe" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.215993 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b050ca-b135-4564-be78-fd2a4efefdbe" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: E1204 18:01:18.216033 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerName="glance-httpd" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216046 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerName="glance-httpd" Dec 04 18:01:18 crc kubenswrapper[4733]: E1204 18:01:18.216061 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e72addf-edb3-4704-be2e-206825774d97" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216076 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e72addf-edb3-4704-be2e-206825774d97" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: E1204 18:01:18.216100 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d45f2b81-a9f9-47bc-8042-c344bffe4dac" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216114 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d45f2b81-a9f9-47bc-8042-c344bffe4dac" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: E1204 18:01:18.216135 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bbdf249-90f9-4685-9418-1ecd4729ac09" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216148 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bbdf249-90f9-4685-9418-1ecd4729ac09" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: E1204 18:01:18.216177 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8436db7-0c0c-4c64-aea2-2266437d9d44" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216190 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8436db7-0c0c-4c64-aea2-2266437d9d44" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216482 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bbdf249-90f9-4685-9418-1ecd4729ac09" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216515 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e72addf-edb3-4704-be2e-206825774d97" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216550 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerName="glance-log" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216571 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d45f2b81-a9f9-47bc-8042-c344bffe4dac" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216583 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b050ca-b135-4564-be78-fd2a4efefdbe" containerName="mariadb-database-create" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216600 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="50ad724d-2883-415e-97f4-636ee63d864c" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216620 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8436db7-0c0c-4c64-aea2-2266437d9d44" containerName="mariadb-account-create-update" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.216645 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" containerName="glance-httpd" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.218340 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.223070 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.223072 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.232922 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.351624 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4f82a6d-dcad-432e-a365-38df92ff4e6e" path="/var/lib/kubelet/pods/f4f82a6d-dcad-432e-a365-38df92ff4e6e/volumes" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.382938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.382989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-logs\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.383034 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.383059 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.383086 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.383114 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9wpc\" (UniqueName: \"kubernetes.io/projected/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-kube-api-access-d9wpc\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.383145 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.383251 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.485842 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-logs\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.485889 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.486018 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.486068 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.486187 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.486236 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9wpc\" (UniqueName: \"kubernetes.io/projected/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-kube-api-access-d9wpc\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.486288 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.486463 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.486901 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.488076 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.488514 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-logs\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.491212 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.491429 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.494333 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.507523 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.521489 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9wpc\" (UniqueName: \"kubernetes.io/projected/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-kube-api-access-d9wpc\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.551761 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " pod="openstack/glance-default-internal-api-0" Dec 04 18:01:18 crc kubenswrapper[4733]: I1204 18:01:18.843098 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.125965 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerStarted","Data":"a798c38f2c9bef4349889f8bf4d6ff55081e2897dd7077bae796799dc1fe21ab"} Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.126266 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.126181 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="ceilometer-central-agent" containerID="cri-o://c095546f6417c9108d455c9a1598286cc75cac04c57dee19b93698265f2c6976" gracePeriod=30 Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.126372 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="sg-core" containerID="cri-o://2edb7919ef80734ff54d234ee326e8e503ecf46f90a48fb930a58d621749bfa2" gracePeriod=30 Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.126401 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="proxy-httpd" containerID="cri-o://a798c38f2c9bef4349889f8bf4d6ff55081e2897dd7077bae796799dc1fe21ab" gracePeriod=30 Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.126376 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="ceilometer-notification-agent" containerID="cri-o://6105deacb60d3518652148ca57b67964adf980319e92bdaaa08f7ca634f506dc" gracePeriod=30 Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.150943 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.630848197 podStartE2EDuration="7.1509208s" podCreationTimestamp="2025-12-04 18:01:12 +0000 UTC" firstStartedPulling="2025-12-04 18:01:13.993959949 +0000 UTC m=+1335.949321005" lastFinishedPulling="2025-12-04 18:01:18.514032562 +0000 UTC m=+1340.469393608" observedRunningTime="2025-12-04 18:01:19.143170086 +0000 UTC m=+1341.098531132" watchObservedRunningTime="2025-12-04 18:01:19.1509208 +0000 UTC m=+1341.106281846" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.348209 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.479654 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t5z7h"] Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.480926 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.484540 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xcbf9" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.484938 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.488986 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.493209 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t5z7h"] Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.611231 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.611920 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-scripts\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.612186 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf9cr\" (UniqueName: \"kubernetes.io/projected/bd664be9-e9ef-4734-8d88-b990f09de1f9-kube-api-access-jf9cr\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.612464 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-config-data\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.715946 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-config-data\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.716039 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.716078 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-scripts\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.716132 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf9cr\" (UniqueName: \"kubernetes.io/projected/bd664be9-e9ef-4734-8d88-b990f09de1f9-kube-api-access-jf9cr\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.721762 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-scripts\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.721817 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.722402 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-config-data\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.734663 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf9cr\" (UniqueName: \"kubernetes.io/projected/bd664be9-e9ef-4734-8d88-b990f09de1f9-kube-api-access-jf9cr\") pod \"nova-cell0-conductor-db-sync-t5z7h\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:19 crc kubenswrapper[4733]: I1204 18:01:19.795324 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.140327 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d","Type":"ContainerStarted","Data":"5fbdddb85e7e76ccd5911f7a283bd09d9eae7deb69d967d864ce48da4e7ac1be"} Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.140875 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d","Type":"ContainerStarted","Data":"554083f376278140ce3ece24f4bafc5e456bf9cc4252f50cdae16d531cec3aeb"} Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.142535 4733 generic.go:334] "Generic (PLEG): container finished" podID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerID="a798c38f2c9bef4349889f8bf4d6ff55081e2897dd7077bae796799dc1fe21ab" exitCode=0 Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.142581 4733 generic.go:334] "Generic (PLEG): container finished" podID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerID="2edb7919ef80734ff54d234ee326e8e503ecf46f90a48fb930a58d621749bfa2" exitCode=2 Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.142591 4733 generic.go:334] "Generic (PLEG): container finished" podID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerID="6105deacb60d3518652148ca57b67964adf980319e92bdaaa08f7ca634f506dc" exitCode=0 Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.142615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerDied","Data":"a798c38f2c9bef4349889f8bf4d6ff55081e2897dd7077bae796799dc1fe21ab"} Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.142644 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerDied","Data":"2edb7919ef80734ff54d234ee326e8e503ecf46f90a48fb930a58d621749bfa2"} Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.142656 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerDied","Data":"6105deacb60d3518652148ca57b67964adf980319e92bdaaa08f7ca634f506dc"} Dec 04 18:01:20 crc kubenswrapper[4733]: I1204 18:01:20.668251 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t5z7h"] Dec 04 18:01:21 crc kubenswrapper[4733]: I1204 18:01:21.153237 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" event={"ID":"bd664be9-e9ef-4734-8d88-b990f09de1f9","Type":"ContainerStarted","Data":"fe708a8316f273d057a233c13154b84fdd7d71dfe1b11543d5d22317ca0a165d"} Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.170310 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d","Type":"ContainerStarted","Data":"6560f5a95ae3f9b652dd82bb257bd795655e85bb32c02059e6da7c32d0b36e95"} Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.184098 4733 generic.go:334] "Generic (PLEG): container finished" podID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerID="c095546f6417c9108d455c9a1598286cc75cac04c57dee19b93698265f2c6976" exitCode=0 Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.184188 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerDied","Data":"c095546f6417c9108d455c9a1598286cc75cac04c57dee19b93698265f2c6976"} Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.193890 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.193873829 podStartE2EDuration="4.193873829s" podCreationTimestamp="2025-12-04 18:01:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:22.186336622 +0000 UTC m=+1344.141697668" watchObservedRunningTime="2025-12-04 18:01:22.193873829 +0000 UTC m=+1344.149234875" Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.923838 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.983431 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-combined-ca-bundle\") pod \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.983473 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-sg-core-conf-yaml\") pod \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.983550 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-log-httpd\") pod \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.983627 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5n88\" (UniqueName: \"kubernetes.io/projected/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-kube-api-access-r5n88\") pod \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.983673 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-config-data\") pod \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.983841 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-scripts\") pod \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.983882 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-run-httpd\") pod \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\" (UID: \"782fb4ec-1220-4f4e-9e74-1ee291ef8c45\") " Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.984578 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "782fb4ec-1220-4f4e-9e74-1ee291ef8c45" (UID: "782fb4ec-1220-4f4e-9e74-1ee291ef8c45"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.985151 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "782fb4ec-1220-4f4e-9e74-1ee291ef8c45" (UID: "782fb4ec-1220-4f4e-9e74-1ee291ef8c45"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.988979 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-kube-api-access-r5n88" (OuterVolumeSpecName: "kube-api-access-r5n88") pod "782fb4ec-1220-4f4e-9e74-1ee291ef8c45" (UID: "782fb4ec-1220-4f4e-9e74-1ee291ef8c45"). InnerVolumeSpecName "kube-api-access-r5n88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:22 crc kubenswrapper[4733]: I1204 18:01:22.989040 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-scripts" (OuterVolumeSpecName: "scripts") pod "782fb4ec-1220-4f4e-9e74-1ee291ef8c45" (UID: "782fb4ec-1220-4f4e-9e74-1ee291ef8c45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.024833 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "782fb4ec-1220-4f4e-9e74-1ee291ef8c45" (UID: "782fb4ec-1220-4f4e-9e74-1ee291ef8c45"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.086076 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.086109 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.086121 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.086137 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.086148 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5n88\" (UniqueName: \"kubernetes.io/projected/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-kube-api-access-r5n88\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.101282 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "782fb4ec-1220-4f4e-9e74-1ee291ef8c45" (UID: "782fb4ec-1220-4f4e-9e74-1ee291ef8c45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.118590 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-config-data" (OuterVolumeSpecName: "config-data") pod "782fb4ec-1220-4f4e-9e74-1ee291ef8c45" (UID: "782fb4ec-1220-4f4e-9e74-1ee291ef8c45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.187204 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.187236 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fb4ec-1220-4f4e-9e74-1ee291ef8c45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.200612 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.202756 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"782fb4ec-1220-4f4e-9e74-1ee291ef8c45","Type":"ContainerDied","Data":"57eb48e9a4d76efb1c5b1071fcebfc100ff53894ee075a47ebb4971ed948727c"} Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.202828 4733 scope.go:117] "RemoveContainer" containerID="a798c38f2c9bef4349889f8bf4d6ff55081e2897dd7077bae796799dc1fe21ab" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.250185 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.250449 4733 scope.go:117] "RemoveContainer" containerID="2edb7919ef80734ff54d234ee326e8e503ecf46f90a48fb930a58d621749bfa2" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.263900 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.273948 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:23 crc kubenswrapper[4733]: E1204 18:01:23.275040 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="ceilometer-central-agent" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.275090 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="ceilometer-central-agent" Dec 04 18:01:23 crc kubenswrapper[4733]: E1204 18:01:23.275114 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="sg-core" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.275125 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="sg-core" Dec 04 18:01:23 crc kubenswrapper[4733]: E1204 18:01:23.275177 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="ceilometer-notification-agent" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.275187 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="ceilometer-notification-agent" Dec 04 18:01:23 crc kubenswrapper[4733]: E1204 18:01:23.275201 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="proxy-httpd" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.275208 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="proxy-httpd" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.275524 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="proxy-httpd" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.275544 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="ceilometer-central-agent" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.275581 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="ceilometer-notification-agent" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.275603 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" containerName="sg-core" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.279628 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.281285 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.282298 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.285606 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.297302 4733 scope.go:117] "RemoveContainer" containerID="6105deacb60d3518652148ca57b67964adf980319e92bdaaa08f7ca634f506dc" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.320416 4733 scope.go:117] "RemoveContainer" containerID="c095546f6417c9108d455c9a1598286cc75cac04c57dee19b93698265f2c6976" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.390028 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-run-httpd\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.390332 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-log-httpd\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.390447 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-config-data\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.390561 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.390729 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.390924 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-scripts\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.391081 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z67s\" (UniqueName: \"kubernetes.io/projected/da79093f-87fe-4488-b06e-cfb50e5b21ba-kube-api-access-7z67s\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.492051 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.492495 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-scripts\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.492603 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z67s\" (UniqueName: \"kubernetes.io/projected/da79093f-87fe-4488-b06e-cfb50e5b21ba-kube-api-access-7z67s\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.492738 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-run-httpd\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.492884 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-log-httpd\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.493001 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-config-data\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.493110 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.495204 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-log-httpd\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.495226 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-run-httpd\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.498745 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-scripts\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.499360 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-config-data\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.500003 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.509609 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.514181 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z67s\" (UniqueName: \"kubernetes.io/projected/da79093f-87fe-4488-b06e-cfb50e5b21ba-kube-api-access-7z67s\") pod \"ceilometer-0\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " pod="openstack/ceilometer-0" Dec 04 18:01:23 crc kubenswrapper[4733]: I1204 18:01:23.595337 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:24 crc kubenswrapper[4733]: I1204 18:01:24.063472 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:24 crc kubenswrapper[4733]: W1204 18:01:24.066266 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda79093f_87fe_4488_b06e_cfb50e5b21ba.slice/crio-3881adafb392d98e5bfc3af55f4cce28054ce11a4d76f883c15d689e590d76a7 WatchSource:0}: Error finding container 3881adafb392d98e5bfc3af55f4cce28054ce11a4d76f883c15d689e590d76a7: Status 404 returned error can't find the container with id 3881adafb392d98e5bfc3af55f4cce28054ce11a4d76f883c15d689e590d76a7 Dec 04 18:01:24 crc kubenswrapper[4733]: I1204 18:01:24.210764 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerStarted","Data":"3881adafb392d98e5bfc3af55f4cce28054ce11a4d76f883c15d689e590d76a7"} Dec 04 18:01:24 crc kubenswrapper[4733]: I1204 18:01:24.346947 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="782fb4ec-1220-4f4e-9e74-1ee291ef8c45" path="/var/lib/kubelet/pods/782fb4ec-1220-4f4e-9e74-1ee291ef8c45/volumes" Dec 04 18:01:24 crc kubenswrapper[4733]: I1204 18:01:24.489216 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 18:01:24 crc kubenswrapper[4733]: I1204 18:01:24.489311 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 18:01:24 crc kubenswrapper[4733]: I1204 18:01:24.532096 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 18:01:24 crc kubenswrapper[4733]: I1204 18:01:24.540484 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 18:01:25 crc kubenswrapper[4733]: I1204 18:01:25.219676 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 18:01:25 crc kubenswrapper[4733]: I1204 18:01:25.220124 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 18:01:27 crc kubenswrapper[4733]: I1204 18:01:27.017206 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:27 crc kubenswrapper[4733]: I1204 18:01:27.043228 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 18:01:27 crc kubenswrapper[4733]: I1204 18:01:27.077642 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 18:01:28 crc kubenswrapper[4733]: I1204 18:01:28.843336 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:28 crc kubenswrapper[4733]: I1204 18:01:28.843700 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:28 crc kubenswrapper[4733]: I1204 18:01:28.877757 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:28 crc kubenswrapper[4733]: I1204 18:01:28.895417 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:29 crc kubenswrapper[4733]: I1204 18:01:29.254060 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:29 crc kubenswrapper[4733]: I1204 18:01:29.254281 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:30 crc kubenswrapper[4733]: I1204 18:01:30.263880 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" event={"ID":"bd664be9-e9ef-4734-8d88-b990f09de1f9","Type":"ContainerStarted","Data":"92cb443d4f09ae45ef70d088a796c54e8df6db7a8ae4d5dd462a4b4e85ed0ade"} Dec 04 18:01:30 crc kubenswrapper[4733]: I1204 18:01:30.268087 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerStarted","Data":"2fc2e35d6a6f29c058ccd9f23da4ef19f9cd7fe5bda6990b82e2a0d7ccb4441b"} Dec 04 18:01:30 crc kubenswrapper[4733]: I1204 18:01:30.291227 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" podStartSLOduration=2.055047933 podStartE2EDuration="11.291206603s" podCreationTimestamp="2025-12-04 18:01:19 +0000 UTC" firstStartedPulling="2025-12-04 18:01:20.673095178 +0000 UTC m=+1342.628456244" lastFinishedPulling="2025-12-04 18:01:29.909253858 +0000 UTC m=+1351.864614914" observedRunningTime="2025-12-04 18:01:30.282374351 +0000 UTC m=+1352.237735437" watchObservedRunningTime="2025-12-04 18:01:30.291206603 +0000 UTC m=+1352.246567649" Dec 04 18:01:31 crc kubenswrapper[4733]: I1204 18:01:31.214292 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:31 crc kubenswrapper[4733]: I1204 18:01:31.217609 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 18:01:32 crc kubenswrapper[4733]: I1204 18:01:32.331463 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerStarted","Data":"2337533eea8111ee991283faa32961642e7538fb2cee5be528239eb8bbbb8938"} Dec 04 18:01:33 crc kubenswrapper[4733]: I1204 18:01:33.342717 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerStarted","Data":"4e3572747d7d94b55c3fcef8c73d11bc9f7f37ca750b4757e9dba7d8277abc9d"} Dec 04 18:01:37 crc kubenswrapper[4733]: I1204 18:01:37.410655 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerStarted","Data":"4c18a94c34e10ed0bfdb145fd2f58a7dc7e54d507e14f001ec439a4b9523d9a9"} Dec 04 18:01:37 crc kubenswrapper[4733]: I1204 18:01:37.411346 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 18:01:37 crc kubenswrapper[4733]: I1204 18:01:37.411003 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="proxy-httpd" containerID="cri-o://4c18a94c34e10ed0bfdb145fd2f58a7dc7e54d507e14f001ec439a4b9523d9a9" gracePeriod=30 Dec 04 18:01:37 crc kubenswrapper[4733]: I1204 18:01:37.411019 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="sg-core" containerID="cri-o://4e3572747d7d94b55c3fcef8c73d11bc9f7f37ca750b4757e9dba7d8277abc9d" gracePeriod=30 Dec 04 18:01:37 crc kubenswrapper[4733]: I1204 18:01:37.411035 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="ceilometer-notification-agent" containerID="cri-o://2337533eea8111ee991283faa32961642e7538fb2cee5be528239eb8bbbb8938" gracePeriod=30 Dec 04 18:01:37 crc kubenswrapper[4733]: I1204 18:01:37.410950 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="ceilometer-central-agent" containerID="cri-o://2fc2e35d6a6f29c058ccd9f23da4ef19f9cd7fe5bda6990b82e2a0d7ccb4441b" gracePeriod=30 Dec 04 18:01:37 crc kubenswrapper[4733]: I1204 18:01:37.434418 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.777697506 podStartE2EDuration="14.434400666s" podCreationTimestamp="2025-12-04 18:01:23 +0000 UTC" firstStartedPulling="2025-12-04 18:01:24.068935948 +0000 UTC m=+1346.024296994" lastFinishedPulling="2025-12-04 18:01:36.725639088 +0000 UTC m=+1358.681000154" observedRunningTime="2025-12-04 18:01:37.432888635 +0000 UTC m=+1359.388249691" watchObservedRunningTime="2025-12-04 18:01:37.434400666 +0000 UTC m=+1359.389761702" Dec 04 18:01:38 crc kubenswrapper[4733]: I1204 18:01:38.422102 4733 generic.go:334] "Generic (PLEG): container finished" podID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerID="4c18a94c34e10ed0bfdb145fd2f58a7dc7e54d507e14f001ec439a4b9523d9a9" exitCode=0 Dec 04 18:01:38 crc kubenswrapper[4733]: I1204 18:01:38.422405 4733 generic.go:334] "Generic (PLEG): container finished" podID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerID="4e3572747d7d94b55c3fcef8c73d11bc9f7f37ca750b4757e9dba7d8277abc9d" exitCode=2 Dec 04 18:01:38 crc kubenswrapper[4733]: I1204 18:01:38.422423 4733 generic.go:334] "Generic (PLEG): container finished" podID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerID="2337533eea8111ee991283faa32961642e7538fb2cee5be528239eb8bbbb8938" exitCode=0 Dec 04 18:01:38 crc kubenswrapper[4733]: I1204 18:01:38.422142 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerDied","Data":"4c18a94c34e10ed0bfdb145fd2f58a7dc7e54d507e14f001ec439a4b9523d9a9"} Dec 04 18:01:38 crc kubenswrapper[4733]: I1204 18:01:38.422466 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerDied","Data":"4e3572747d7d94b55c3fcef8c73d11bc9f7f37ca750b4757e9dba7d8277abc9d"} Dec 04 18:01:38 crc kubenswrapper[4733]: I1204 18:01:38.422483 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerDied","Data":"2337533eea8111ee991283faa32961642e7538fb2cee5be528239eb8bbbb8938"} Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.464244 4733 generic.go:334] "Generic (PLEG): container finished" podID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerID="2fc2e35d6a6f29c058ccd9f23da4ef19f9cd7fe5bda6990b82e2a0d7ccb4441b" exitCode=0 Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.464574 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerDied","Data":"2fc2e35d6a6f29c058ccd9f23da4ef19f9cd7fe5bda6990b82e2a0d7ccb4441b"} Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.590141 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.751097 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-config-data\") pod \"da79093f-87fe-4488-b06e-cfb50e5b21ba\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.751200 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-run-httpd\") pod \"da79093f-87fe-4488-b06e-cfb50e5b21ba\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.751284 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-combined-ca-bundle\") pod \"da79093f-87fe-4488-b06e-cfb50e5b21ba\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.751348 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-sg-core-conf-yaml\") pod \"da79093f-87fe-4488-b06e-cfb50e5b21ba\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.751371 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-scripts\") pod \"da79093f-87fe-4488-b06e-cfb50e5b21ba\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.751391 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-log-httpd\") pod \"da79093f-87fe-4488-b06e-cfb50e5b21ba\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.751458 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z67s\" (UniqueName: \"kubernetes.io/projected/da79093f-87fe-4488-b06e-cfb50e5b21ba-kube-api-access-7z67s\") pod \"da79093f-87fe-4488-b06e-cfb50e5b21ba\" (UID: \"da79093f-87fe-4488-b06e-cfb50e5b21ba\") " Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.753860 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "da79093f-87fe-4488-b06e-cfb50e5b21ba" (UID: "da79093f-87fe-4488-b06e-cfb50e5b21ba"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.754211 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "da79093f-87fe-4488-b06e-cfb50e5b21ba" (UID: "da79093f-87fe-4488-b06e-cfb50e5b21ba"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.758287 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-scripts" (OuterVolumeSpecName: "scripts") pod "da79093f-87fe-4488-b06e-cfb50e5b21ba" (UID: "da79093f-87fe-4488-b06e-cfb50e5b21ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.758486 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da79093f-87fe-4488-b06e-cfb50e5b21ba-kube-api-access-7z67s" (OuterVolumeSpecName: "kube-api-access-7z67s") pod "da79093f-87fe-4488-b06e-cfb50e5b21ba" (UID: "da79093f-87fe-4488-b06e-cfb50e5b21ba"). InnerVolumeSpecName "kube-api-access-7z67s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.805261 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "da79093f-87fe-4488-b06e-cfb50e5b21ba" (UID: "da79093f-87fe-4488-b06e-cfb50e5b21ba"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.839006 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da79093f-87fe-4488-b06e-cfb50e5b21ba" (UID: "da79093f-87fe-4488-b06e-cfb50e5b21ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.853582 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.853612 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.853625 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.853636 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.853647 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da79093f-87fe-4488-b06e-cfb50e5b21ba-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.853658 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z67s\" (UniqueName: \"kubernetes.io/projected/da79093f-87fe-4488-b06e-cfb50e5b21ba-kube-api-access-7z67s\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.890313 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-config-data" (OuterVolumeSpecName: "config-data") pod "da79093f-87fe-4488-b06e-cfb50e5b21ba" (UID: "da79093f-87fe-4488-b06e-cfb50e5b21ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:42 crc kubenswrapper[4733]: I1204 18:01:42.956377 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da79093f-87fe-4488-b06e-cfb50e5b21ba-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.476621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da79093f-87fe-4488-b06e-cfb50e5b21ba","Type":"ContainerDied","Data":"3881adafb392d98e5bfc3af55f4cce28054ce11a4d76f883c15d689e590d76a7"} Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.476696 4733 scope.go:117] "RemoveContainer" containerID="4c18a94c34e10ed0bfdb145fd2f58a7dc7e54d507e14f001ec439a4b9523d9a9" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.476696 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.516340 4733 scope.go:117] "RemoveContainer" containerID="4e3572747d7d94b55c3fcef8c73d11bc9f7f37ca750b4757e9dba7d8277abc9d" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.516626 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.527624 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.543155 4733 scope.go:117] "RemoveContainer" containerID="2337533eea8111ee991283faa32961642e7538fb2cee5be528239eb8bbbb8938" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.544602 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:43 crc kubenswrapper[4733]: E1204 18:01:43.545157 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="ceilometer-notification-agent" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.545174 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="ceilometer-notification-agent" Dec 04 18:01:43 crc kubenswrapper[4733]: E1204 18:01:43.545193 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="proxy-httpd" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.545199 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="proxy-httpd" Dec 04 18:01:43 crc kubenswrapper[4733]: E1204 18:01:43.545209 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="ceilometer-central-agent" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.545217 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="ceilometer-central-agent" Dec 04 18:01:43 crc kubenswrapper[4733]: E1204 18:01:43.545240 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="sg-core" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.545246 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="sg-core" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.545445 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="proxy-httpd" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.545459 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="sg-core" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.545466 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="ceilometer-central-agent" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.545475 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" containerName="ceilometer-notification-agent" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.547086 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.549511 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.549556 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.566538 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.574232 4733 scope.go:117] "RemoveContainer" containerID="2fc2e35d6a6f29c058ccd9f23da4ef19f9cd7fe5bda6990b82e2a0d7ccb4441b" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.670904 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-log-httpd\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.671005 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.671094 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wgcx\" (UniqueName: \"kubernetes.io/projected/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-kube-api-access-8wgcx\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.671173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.671239 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-config-data\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.671260 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-run-httpd\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.671286 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-scripts\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.772870 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.772947 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-config-data\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.772969 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-run-httpd\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.772989 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-scripts\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.773025 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-log-httpd\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.773055 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.773084 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wgcx\" (UniqueName: \"kubernetes.io/projected/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-kube-api-access-8wgcx\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.774164 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-log-httpd\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.774567 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-run-httpd\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.778146 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.779192 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-scripts\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.783147 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.792467 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-config-data\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.796701 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wgcx\" (UniqueName: \"kubernetes.io/projected/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-kube-api-access-8wgcx\") pod \"ceilometer-0\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " pod="openstack/ceilometer-0" Dec 04 18:01:43 crc kubenswrapper[4733]: I1204 18:01:43.892671 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:44 crc kubenswrapper[4733]: I1204 18:01:44.359006 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da79093f-87fe-4488-b06e-cfb50e5b21ba" path="/var/lib/kubelet/pods/da79093f-87fe-4488-b06e-cfb50e5b21ba/volumes" Dec 04 18:01:44 crc kubenswrapper[4733]: I1204 18:01:44.389249 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:44 crc kubenswrapper[4733]: I1204 18:01:44.490089 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerStarted","Data":"1b901b52cc0202a770df290cec3c5903308feb833e5e60f19b3f8ddfeb117245"} Dec 04 18:01:45 crc kubenswrapper[4733]: I1204 18:01:45.361934 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:01:45 crc kubenswrapper[4733]: I1204 18:01:45.362257 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:01:45 crc kubenswrapper[4733]: I1204 18:01:45.520090 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerStarted","Data":"4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516"} Dec 04 18:01:45 crc kubenswrapper[4733]: I1204 18:01:45.521449 4733 generic.go:334] "Generic (PLEG): container finished" podID="bd664be9-e9ef-4734-8d88-b990f09de1f9" containerID="92cb443d4f09ae45ef70d088a796c54e8df6db7a8ae4d5dd462a4b4e85ed0ade" exitCode=0 Dec 04 18:01:45 crc kubenswrapper[4733]: I1204 18:01:45.521478 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" event={"ID":"bd664be9-e9ef-4734-8d88-b990f09de1f9","Type":"ContainerDied","Data":"92cb443d4f09ae45ef70d088a796c54e8df6db7a8ae4d5dd462a4b4e85ed0ade"} Dec 04 18:01:46 crc kubenswrapper[4733]: I1204 18:01:46.532541 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerStarted","Data":"42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a"} Dec 04 18:01:46 crc kubenswrapper[4733]: I1204 18:01:46.965102 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.062637 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-config-data\") pod \"bd664be9-e9ef-4734-8d88-b990f09de1f9\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.062741 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf9cr\" (UniqueName: \"kubernetes.io/projected/bd664be9-e9ef-4734-8d88-b990f09de1f9-kube-api-access-jf9cr\") pod \"bd664be9-e9ef-4734-8d88-b990f09de1f9\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.062838 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-scripts\") pod \"bd664be9-e9ef-4734-8d88-b990f09de1f9\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.062857 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-combined-ca-bundle\") pod \"bd664be9-e9ef-4734-8d88-b990f09de1f9\" (UID: \"bd664be9-e9ef-4734-8d88-b990f09de1f9\") " Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.070150 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd664be9-e9ef-4734-8d88-b990f09de1f9-kube-api-access-jf9cr" (OuterVolumeSpecName: "kube-api-access-jf9cr") pod "bd664be9-e9ef-4734-8d88-b990f09de1f9" (UID: "bd664be9-e9ef-4734-8d88-b990f09de1f9"). InnerVolumeSpecName "kube-api-access-jf9cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.077991 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-scripts" (OuterVolumeSpecName: "scripts") pod "bd664be9-e9ef-4734-8d88-b990f09de1f9" (UID: "bd664be9-e9ef-4734-8d88-b990f09de1f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.097000 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd664be9-e9ef-4734-8d88-b990f09de1f9" (UID: "bd664be9-e9ef-4734-8d88-b990f09de1f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.106130 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-config-data" (OuterVolumeSpecName: "config-data") pod "bd664be9-e9ef-4734-8d88-b990f09de1f9" (UID: "bd664be9-e9ef-4734-8d88-b990f09de1f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.164930 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf9cr\" (UniqueName: \"kubernetes.io/projected/bd664be9-e9ef-4734-8d88-b990f09de1f9-kube-api-access-jf9cr\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.164968 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.164985 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.164998 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd664be9-e9ef-4734-8d88-b990f09de1f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.442869 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.541934 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerStarted","Data":"1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d"} Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.543401 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" event={"ID":"bd664be9-e9ef-4734-8d88-b990f09de1f9","Type":"ContainerDied","Data":"fe708a8316f273d057a233c13154b84fdd7d71dfe1b11543d5d22317ca0a165d"} Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.543428 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe708a8316f273d057a233c13154b84fdd7d71dfe1b11543d5d22317ca0a165d" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.543482 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-t5z7h" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.691548 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 18:01:47 crc kubenswrapper[4733]: E1204 18:01:47.692272 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd664be9-e9ef-4734-8d88-b990f09de1f9" containerName="nova-cell0-conductor-db-sync" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.692295 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd664be9-e9ef-4734-8d88-b990f09de1f9" containerName="nova-cell0-conductor-db-sync" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.692484 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd664be9-e9ef-4734-8d88-b990f09de1f9" containerName="nova-cell0-conductor-db-sync" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.693365 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.695785 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xcbf9" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.695965 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.723712 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.877157 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.877266 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.877352 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d642r\" (UniqueName: \"kubernetes.io/projected/5eecf779-672e-4469-a134-9ecb0ef1bdde-kube-api-access-d642r\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.979257 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.979345 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.979513 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d642r\" (UniqueName: \"kubernetes.io/projected/5eecf779-672e-4469-a134-9ecb0ef1bdde-kube-api-access-d642r\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:47 crc kubenswrapper[4733]: I1204 18:01:47.986496 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.000015 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.005524 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d642r\" (UniqueName: \"kubernetes.io/projected/5eecf779-672e-4469-a134-9ecb0ef1bdde-kube-api-access-d642r\") pod \"nova-cell0-conductor-0\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.059548 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.518886 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 18:01:48 crc kubenswrapper[4733]: W1204 18:01:48.526168 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5eecf779_672e_4469_a134_9ecb0ef1bdde.slice/crio-7b4e7fd78c569c81d6798502728ecf4b61b99db55f050463555b435ea0c256e5 WatchSource:0}: Error finding container 7b4e7fd78c569c81d6798502728ecf4b61b99db55f050463555b435ea0c256e5: Status 404 returned error can't find the container with id 7b4e7fd78c569c81d6798502728ecf4b61b99db55f050463555b435ea0c256e5 Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.553609 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5eecf779-672e-4469-a134-9ecb0ef1bdde","Type":"ContainerStarted","Data":"7b4e7fd78c569c81d6798502728ecf4b61b99db55f050463555b435ea0c256e5"} Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.557725 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerStarted","Data":"602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b"} Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.558179 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="ceilometer-central-agent" containerID="cri-o://4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516" gracePeriod=30 Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.558419 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.558895 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="proxy-httpd" containerID="cri-o://602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b" gracePeriod=30 Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.558945 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="sg-core" containerID="cri-o://1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d" gracePeriod=30 Dec 04 18:01:48 crc kubenswrapper[4733]: I1204 18:01:48.558907 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="ceilometer-notification-agent" containerID="cri-o://42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a" gracePeriod=30 Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.574271 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5eecf779-672e-4469-a134-9ecb0ef1bdde","Type":"ContainerStarted","Data":"96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b"} Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.574872 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.579541 4733 generic.go:334] "Generic (PLEG): container finished" podID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerID="602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b" exitCode=0 Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.579570 4733 generic.go:334] "Generic (PLEG): container finished" podID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerID="1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d" exitCode=2 Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.579578 4733 generic.go:334] "Generic (PLEG): container finished" podID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerID="42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a" exitCode=0 Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.579600 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerDied","Data":"602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b"} Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.579625 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerDied","Data":"1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d"} Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.579636 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerDied","Data":"42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a"} Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.604764 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.604742038 podStartE2EDuration="2.604742038s" podCreationTimestamp="2025-12-04 18:01:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:49.60295176 +0000 UTC m=+1371.558312856" watchObservedRunningTime="2025-12-04 18:01:49.604742038 +0000 UTC m=+1371.560103094" Dec 04 18:01:49 crc kubenswrapper[4733]: I1204 18:01:49.611344 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.845150911 podStartE2EDuration="6.611306992s" podCreationTimestamp="2025-12-04 18:01:43 +0000 UTC" firstStartedPulling="2025-12-04 18:01:44.399076124 +0000 UTC m=+1366.354437170" lastFinishedPulling="2025-12-04 18:01:48.165232175 +0000 UTC m=+1370.120593251" observedRunningTime="2025-12-04 18:01:48.585005261 +0000 UTC m=+1370.540366307" watchObservedRunningTime="2025-12-04 18:01:49.611306992 +0000 UTC m=+1371.566668048" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.461307 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.564457 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-scripts\") pod \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.564593 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-combined-ca-bundle\") pod \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.564661 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-config-data\") pod \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.564685 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wgcx\" (UniqueName: \"kubernetes.io/projected/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-kube-api-access-8wgcx\") pod \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.564742 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-log-httpd\") pod \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.564770 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-sg-core-conf-yaml\") pod \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.564855 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-run-httpd\") pod \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\" (UID: \"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a\") " Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.565297 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" (UID: "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.565469 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" (UID: "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.570200 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-scripts" (OuterVolumeSpecName: "scripts") pod "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" (UID: "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.572123 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-kube-api-access-8wgcx" (OuterVolumeSpecName: "kube-api-access-8wgcx") pod "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" (UID: "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a"). InnerVolumeSpecName "kube-api-access-8wgcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.598642 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" (UID: "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.635492 4733 generic.go:334] "Generic (PLEG): container finished" podID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerID="4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516" exitCode=0 Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.635570 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerDied","Data":"4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516"} Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.635578 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.635612 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a","Type":"ContainerDied","Data":"1b901b52cc0202a770df290cec3c5903308feb833e5e60f19b3f8ddfeb117245"} Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.635655 4733 scope.go:117] "RemoveContainer" containerID="602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.654504 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" (UID: "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.666526 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.666560 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.666570 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wgcx\" (UniqueName: \"kubernetes.io/projected/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-kube-api-access-8wgcx\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.666579 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.666588 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.666598 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.668861 4733 scope.go:117] "RemoveContainer" containerID="1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.687622 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-config-data" (OuterVolumeSpecName: "config-data") pod "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" (UID: "ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.688786 4733 scope.go:117] "RemoveContainer" containerID="42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.710245 4733 scope.go:117] "RemoveContainer" containerID="4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.739955 4733 scope.go:117] "RemoveContainer" containerID="602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b" Dec 04 18:01:52 crc kubenswrapper[4733]: E1204 18:01:52.740539 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b\": container with ID starting with 602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b not found: ID does not exist" containerID="602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.740591 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b"} err="failed to get container status \"602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b\": rpc error: code = NotFound desc = could not find container \"602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b\": container with ID starting with 602324b2185a03aef2e956c75218678fa44226a98e9e80ed5e11d1a777fb7d1b not found: ID does not exist" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.740617 4733 scope.go:117] "RemoveContainer" containerID="1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d" Dec 04 18:01:52 crc kubenswrapper[4733]: E1204 18:01:52.741249 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d\": container with ID starting with 1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d not found: ID does not exist" containerID="1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.741348 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d"} err="failed to get container status \"1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d\": rpc error: code = NotFound desc = could not find container \"1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d\": container with ID starting with 1ef026b7e5023c3606f1a6a422075c09f221d3b28fd76781c1c03f3775710f0d not found: ID does not exist" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.741435 4733 scope.go:117] "RemoveContainer" containerID="42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a" Dec 04 18:01:52 crc kubenswrapper[4733]: E1204 18:01:52.742143 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a\": container with ID starting with 42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a not found: ID does not exist" containerID="42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.742175 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a"} err="failed to get container status \"42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a\": rpc error: code = NotFound desc = could not find container \"42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a\": container with ID starting with 42a049e1f83e26c64d134de6a7c7abdda90caa8265838ce879db20e58459e70a not found: ID does not exist" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.742218 4733 scope.go:117] "RemoveContainer" containerID="4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516" Dec 04 18:01:52 crc kubenswrapper[4733]: E1204 18:01:52.742471 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516\": container with ID starting with 4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516 not found: ID does not exist" containerID="4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.742550 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516"} err="failed to get container status \"4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516\": rpc error: code = NotFound desc = could not find container \"4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516\": container with ID starting with 4a119b22611a95e2fa1f9c2188f818ccdc39ce129c255a6612cc1de052156516 not found: ID does not exist" Dec 04 18:01:52 crc kubenswrapper[4733]: I1204 18:01:52.769032 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.010711 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.026054 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.039779 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: E1204 18:01:53.040317 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="sg-core" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.040338 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="sg-core" Dec 04 18:01:53 crc kubenswrapper[4733]: E1204 18:01:53.040367 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="ceilometer-notification-agent" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.040373 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="ceilometer-notification-agent" Dec 04 18:01:53 crc kubenswrapper[4733]: E1204 18:01:53.040389 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="proxy-httpd" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.040395 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="proxy-httpd" Dec 04 18:01:53 crc kubenswrapper[4733]: E1204 18:01:53.040404 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="ceilometer-central-agent" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.040410 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="ceilometer-central-agent" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.040594 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="ceilometer-notification-agent" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.040605 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="ceilometer-central-agent" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.040616 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="sg-core" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.040629 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" containerName="proxy-httpd" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.042185 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.046620 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.046840 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.047030 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.097712 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.176073 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.176169 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-run-httpd\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.176224 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7zjh\" (UniqueName: \"kubernetes.io/projected/32c62974-76c1-4ce1-8c63-0c231ef8e38c-kube-api-access-f7zjh\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.176249 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-scripts\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.176269 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-log-httpd\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.176313 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-config-data\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.176346 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.277902 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-config-data\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.278024 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.278092 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.278212 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-run-httpd\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.278287 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7zjh\" (UniqueName: \"kubernetes.io/projected/32c62974-76c1-4ce1-8c63-0c231ef8e38c-kube-api-access-f7zjh\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.278345 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-scripts\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.278373 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-log-httpd\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.278864 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-run-httpd\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.279049 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-log-httpd\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.282853 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.283321 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-scripts\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.284054 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-config-data\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.289085 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.306658 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7zjh\" (UniqueName: \"kubernetes.io/projected/32c62974-76c1-4ce1-8c63-0c231ef8e38c-kube-api-access-f7zjh\") pod \"ceilometer-0\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.371543 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.586137 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xv4z2"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.587503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.592390 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.592712 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.596711 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xv4z2"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.695590 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-config-data\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.695691 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.695730 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-scripts\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.695764 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5czz5\" (UniqueName: \"kubernetes.io/projected/553ecb2b-6910-4f64-ac13-93fa1afe5137-kube-api-access-5czz5\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.734942 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.741454 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.747140 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.765039 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.799901 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.799963 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-scripts\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.800005 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5czz5\" (UniqueName: \"kubernetes.io/projected/553ecb2b-6910-4f64-ac13-93fa1afe5137-kube-api-access-5czz5\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.800055 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-config-data\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.811770 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.811846 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-scripts\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.817324 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-config-data\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.834226 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.835927 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.842015 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.848529 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5czz5\" (UniqueName: \"kubernetes.io/projected/553ecb2b-6910-4f64-ac13-93fa1afe5137-kube-api-access-5czz5\") pod \"nova-cell0-cell-mapping-xv4z2\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.880261 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.907223 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-logs\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.907657 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.907806 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw6zx\" (UniqueName: \"kubernetes.io/projected/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-kube-api-access-gw6zx\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.907953 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-config-data\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.918111 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.933837 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.937005 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.946021 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.946402 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 18:01:53 crc kubenswrapper[4733]: I1204 18:01:53.968844 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.009949 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc92j\" (UniqueName: \"kubernetes.io/projected/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-kube-api-access-kc92j\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.009993 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-config-data\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.010045 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-logs\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.010076 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.010155 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.010199 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw6zx\" (UniqueName: \"kubernetes.io/projected/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-kube-api-access-gw6zx\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.010239 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-logs\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.010261 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-config-data\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.010431 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-logs\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.012835 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.014119 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.015969 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.019646 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.024253 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-qbzmf"] Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.026570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.027434 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw6zx\" (UniqueName: \"kubernetes.io/projected/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-kube-api-access-gw6zx\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.027449 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-config-data\") pod \"nova-api-0\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.055460 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.065227 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.075047 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-qbzmf"] Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.113469 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc92j\" (UniqueName: \"kubernetes.io/projected/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-kube-api-access-kc92j\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.113993 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-config-data\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.114071 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-config-data\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.114160 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.114239 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.114338 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjlp4\" (UniqueName: \"kubernetes.io/projected/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-kube-api-access-mjlp4\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.114459 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.114485 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6bgq\" (UniqueName: \"kubernetes.io/projected/f01545db-3507-4b75-ba11-b37afd1770d7-kube-api-access-l6bgq\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.114562 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.114620 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-logs\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.115267 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-logs\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.142517 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-config-data\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.151897 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc92j\" (UniqueName: \"kubernetes.io/projected/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-kube-api-access-kc92j\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.153234 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.165237 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217399 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217476 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217536 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-config\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217586 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217608 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-config-data\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217630 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217658 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217677 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217742 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjlp4\" (UniqueName: \"kubernetes.io/projected/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-kube-api-access-mjlp4\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.217763 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f6bt\" (UniqueName: \"kubernetes.io/projected/16c800b6-4095-4676-ac1a-2125b3b40031-kube-api-access-9f6bt\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.218131 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.218187 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6bgq\" (UniqueName: \"kubernetes.io/projected/f01545db-3507-4b75-ba11-b37afd1770d7-kube-api-access-l6bgq\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.222250 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.229763 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-config-data\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.232695 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.245485 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.271478 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjlp4\" (UniqueName: \"kubernetes.io/projected/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-kube-api-access-mjlp4\") pod \"nova-cell1-novncproxy-0\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.273654 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6bgq\" (UniqueName: \"kubernetes.io/projected/f01545db-3507-4b75-ba11-b37afd1770d7-kube-api-access-l6bgq\") pod \"nova-scheduler-0\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.279922 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.328541 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f6bt\" (UniqueName: \"kubernetes.io/projected/16c800b6-4095-4676-ac1a-2125b3b40031-kube-api-access-9f6bt\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.328615 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.328659 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-config\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.328702 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.328737 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.328756 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.329611 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-swift-storage-0\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.329686 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-sb\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.329868 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-nb\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.330424 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-svc\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.330668 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-config\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.353865 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f6bt\" (UniqueName: \"kubernetes.io/projected/16c800b6-4095-4676-ac1a-2125b3b40031-kube-api-access-9f6bt\") pod \"dnsmasq-dns-7d75688ddc-qbzmf\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.361112 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a" path="/var/lib/kubelet/pods/ffb1d2ac-0b58-4198-8a5a-41ae06b0d19a/volumes" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.366858 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.398275 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.539358 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xv4z2"] Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.698367 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xv4z2" event={"ID":"553ecb2b-6910-4f64-ac13-93fa1afe5137","Type":"ContainerStarted","Data":"278ce4720c576463280f17cef3ba35d8061f00ea087467269cfbe9137c96787d"} Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.712710 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerStarted","Data":"9197e904a60e2e6b62980c11341a1c87487bcd838cfc655374722c6eb66a56ee"} Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.741709 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:01:54 crc kubenswrapper[4733]: W1204 18:01:54.793327 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a98ae96_6a22_4fcb_8bb5_e7bd45f78b6d.slice/crio-5adb5f7f07512cf6b3b88cf007522a9b618d4c80fdf3d8b97d98eb11296215c4 WatchSource:0}: Error finding container 5adb5f7f07512cf6b3b88cf007522a9b618d4c80fdf3d8b97d98eb11296215c4: Status 404 returned error can't find the container with id 5adb5f7f07512cf6b3b88cf007522a9b618d4c80fdf3d8b97d98eb11296215c4 Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.895650 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:01:54 crc kubenswrapper[4733]: W1204 18:01:54.902430 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf01545db_3507_4b75_ba11_b37afd1770d7.slice/crio-5af90d347b9c133c79931454bf8438a50241b223d51c9c7a5a04f5de2f51219c WatchSource:0}: Error finding container 5af90d347b9c133c79931454bf8438a50241b223d51c9c7a5a04f5de2f51219c: Status 404 returned error can't find the container with id 5af90d347b9c133c79931454bf8438a50241b223d51c9c7a5a04f5de2f51219c Dec 04 18:01:54 crc kubenswrapper[4733]: I1204 18:01:54.905636 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.023508 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:01:55 crc kubenswrapper[4733]: W1204 18:01:55.034542 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76ab1aa1_894a_49b4_a1be_8b91429bf2ed.slice/crio-680b7bc2eebd0baa9426d98543482e1111073c5cb7b2f27cd2b0fba415bf199a WatchSource:0}: Error finding container 680b7bc2eebd0baa9426d98543482e1111073c5cb7b2f27cd2b0fba415bf199a: Status 404 returned error can't find the container with id 680b7bc2eebd0baa9426d98543482e1111073c5cb7b2f27cd2b0fba415bf199a Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.035987 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8fxb9"] Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.037042 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.040184 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.040398 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.057342 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8fxb9"] Dec 04 18:01:55 crc kubenswrapper[4733]: W1204 18:01:55.121817 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16c800b6_4095_4676_ac1a_2125b3b40031.slice/crio-94f93239559b64b542a844ee5a5a4a77b9857fc47928b4c69e8be150f42b9ed0 WatchSource:0}: Error finding container 94f93239559b64b542a844ee5a5a4a77b9857fc47928b4c69e8be150f42b9ed0: Status 404 returned error can't find the container with id 94f93239559b64b542a844ee5a5a4a77b9857fc47928b4c69e8be150f42b9ed0 Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.124386 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-qbzmf"] Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.148467 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.148535 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4h4b\" (UniqueName: \"kubernetes.io/projected/18512067-9d64-4796-9bc1-a3d83ef5b13e-kube-api-access-v4h4b\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.148561 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-config-data\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.148671 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-scripts\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.250533 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-scripts\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.250815 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.250936 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4h4b\" (UniqueName: \"kubernetes.io/projected/18512067-9d64-4796-9bc1-a3d83ef5b13e-kube-api-access-v4h4b\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.251029 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-config-data\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.254292 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-scripts\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.254774 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.255086 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-config-data\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.274369 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4h4b\" (UniqueName: \"kubernetes.io/projected/18512067-9d64-4796-9bc1-a3d83ef5b13e-kube-api-access-v4h4b\") pod \"nova-cell1-conductor-db-sync-8fxb9\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.360144 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.739218 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xv4z2" event={"ID":"553ecb2b-6910-4f64-ac13-93fa1afe5137","Type":"ContainerStarted","Data":"b5cab691f053ad9777f8ae8b5b15035ab043220183fb8752c859b471c466a3d8"} Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.754569 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xv4z2" podStartSLOduration=2.75455058 podStartE2EDuration="2.75455058s" podCreationTimestamp="2025-12-04 18:01:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:55.753621156 +0000 UTC m=+1377.708982202" watchObservedRunningTime="2025-12-04 18:01:55.75455058 +0000 UTC m=+1377.709911626" Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.760226 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerStarted","Data":"3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112"} Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.764686 4733 generic.go:334] "Generic (PLEG): container finished" podID="16c800b6-4095-4676-ac1a-2125b3b40031" containerID="9063871b77f8aabc81ae5aa75dd5adc6cab49b2defefd7449f490514978e0928" exitCode=0 Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.765332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" event={"ID":"16c800b6-4095-4676-ac1a-2125b3b40031","Type":"ContainerDied","Data":"9063871b77f8aabc81ae5aa75dd5adc6cab49b2defefd7449f490514978e0928"} Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.771374 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" event={"ID":"16c800b6-4095-4676-ac1a-2125b3b40031","Type":"ContainerStarted","Data":"94f93239559b64b542a844ee5a5a4a77b9857fc47928b4c69e8be150f42b9ed0"} Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.771424 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f01545db-3507-4b75-ba11-b37afd1770d7","Type":"ContainerStarted","Data":"5af90d347b9c133c79931454bf8438a50241b223d51c9c7a5a04f5de2f51219c"} Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.771545 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"76ab1aa1-894a-49b4-a1be-8b91429bf2ed","Type":"ContainerStarted","Data":"680b7bc2eebd0baa9426d98543482e1111073c5cb7b2f27cd2b0fba415bf199a"} Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.781950 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d","Type":"ContainerStarted","Data":"5adb5f7f07512cf6b3b88cf007522a9b618d4c80fdf3d8b97d98eb11296215c4"} Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.791085 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0963c49e-2e7a-4ae8-b6d7-55f1926d7476","Type":"ContainerStarted","Data":"662c9edeb67568f3f2b5b3797cb11b0e3e318d16977d0958dc8fec6588c69474"} Dec 04 18:01:55 crc kubenswrapper[4733]: I1204 18:01:55.849137 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8fxb9"] Dec 04 18:01:55 crc kubenswrapper[4733]: W1204 18:01:55.857540 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18512067_9d64_4796_9bc1_a3d83ef5b13e.slice/crio-9572c9403d011501520ebd9ff21553a210af80301974d5f916fc484ed7bd7348 WatchSource:0}: Error finding container 9572c9403d011501520ebd9ff21553a210af80301974d5f916fc484ed7bd7348: Status 404 returned error can't find the container with id 9572c9403d011501520ebd9ff21553a210af80301974d5f916fc484ed7bd7348 Dec 04 18:01:56 crc kubenswrapper[4733]: I1204 18:01:56.806269 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" event={"ID":"18512067-9d64-4796-9bc1-a3d83ef5b13e","Type":"ContainerStarted","Data":"9faff0680efdb357f0f7ab94276aac3340ec7dc913f22f91f8b0cae7b5683fd3"} Dec 04 18:01:56 crc kubenswrapper[4733]: I1204 18:01:56.806596 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" event={"ID":"18512067-9d64-4796-9bc1-a3d83ef5b13e","Type":"ContainerStarted","Data":"9572c9403d011501520ebd9ff21553a210af80301974d5f916fc484ed7bd7348"} Dec 04 18:01:56 crc kubenswrapper[4733]: I1204 18:01:56.809696 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerStarted","Data":"f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589"} Dec 04 18:01:56 crc kubenswrapper[4733]: I1204 18:01:56.812249 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" event={"ID":"16c800b6-4095-4676-ac1a-2125b3b40031","Type":"ContainerStarted","Data":"f32eabfab4f0ea2f0f6ea5778bd2054fd4fbd57466933574641025b23450044f"} Dec 04 18:01:56 crc kubenswrapper[4733]: I1204 18:01:56.822479 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" podStartSLOduration=1.822457677 podStartE2EDuration="1.822457677s" podCreationTimestamp="2025-12-04 18:01:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:56.822256411 +0000 UTC m=+1378.777617467" watchObservedRunningTime="2025-12-04 18:01:56.822457677 +0000 UTC m=+1378.777818733" Dec 04 18:01:56 crc kubenswrapper[4733]: I1204 18:01:56.848930 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" podStartSLOduration=3.848911209 podStartE2EDuration="3.848911209s" podCreationTimestamp="2025-12-04 18:01:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:01:56.84555727 +0000 UTC m=+1378.800918336" watchObservedRunningTime="2025-12-04 18:01:56.848911209 +0000 UTC m=+1378.804272275" Dec 04 18:01:57 crc kubenswrapper[4733]: I1204 18:01:57.821728 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:01:57 crc kubenswrapper[4733]: I1204 18:01:57.915503 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:01:57 crc kubenswrapper[4733]: I1204 18:01:57.924787 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.842715 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d","Type":"ContainerStarted","Data":"6602dc5879168dcaaa99620ea72395421ca373d0f88940c5eada9244c55900a0"} Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.843358 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d","Type":"ContainerStarted","Data":"a663e1bc5727c5ca185760cb484948770d7432190c4cda6e4378bca5a497f3b1"} Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.847066 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0963c49e-2e7a-4ae8-b6d7-55f1926d7476","Type":"ContainerStarted","Data":"5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2"} Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.847109 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0963c49e-2e7a-4ae8-b6d7-55f1926d7476","Type":"ContainerStarted","Data":"32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974"} Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.847472 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerName="nova-metadata-metadata" containerID="cri-o://5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2" gracePeriod=30 Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.847425 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerName="nova-metadata-log" containerID="cri-o://32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974" gracePeriod=30 Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.850267 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerStarted","Data":"82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080"} Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.853459 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f01545db-3507-4b75-ba11-b37afd1770d7","Type":"ContainerStarted","Data":"f9319c0a859fc9df0172277a65be92b8bdf0941cf298061e0067aeb25c1626f3"} Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.855385 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"76ab1aa1-894a-49b4-a1be-8b91429bf2ed","Type":"ContainerStarted","Data":"47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3"} Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.855544 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="76ab1aa1-894a-49b4-a1be-8b91429bf2ed" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3" gracePeriod=30 Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.866694 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.44265761 podStartE2EDuration="6.866671258s" podCreationTimestamp="2025-12-04 18:01:53 +0000 UTC" firstStartedPulling="2025-12-04 18:01:54.7952834 +0000 UTC m=+1376.750644446" lastFinishedPulling="2025-12-04 18:01:59.219297048 +0000 UTC m=+1381.174658094" observedRunningTime="2025-12-04 18:01:59.860207496 +0000 UTC m=+1381.815568542" watchObservedRunningTime="2025-12-04 18:01:59.866671258 +0000 UTC m=+1381.822032304" Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.885305 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.703028222 podStartE2EDuration="6.885287782s" podCreationTimestamp="2025-12-04 18:01:53 +0000 UTC" firstStartedPulling="2025-12-04 18:01:55.037633044 +0000 UTC m=+1376.992994090" lastFinishedPulling="2025-12-04 18:01:59.219892604 +0000 UTC m=+1381.175253650" observedRunningTime="2025-12-04 18:01:59.879005286 +0000 UTC m=+1381.834366352" watchObservedRunningTime="2025-12-04 18:01:59.885287782 +0000 UTC m=+1381.840648828" Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.908363 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.595136187 podStartE2EDuration="6.908344694s" podCreationTimestamp="2025-12-04 18:01:53 +0000 UTC" firstStartedPulling="2025-12-04 18:01:54.911782413 +0000 UTC m=+1376.867143459" lastFinishedPulling="2025-12-04 18:01:59.22499092 +0000 UTC m=+1381.180351966" observedRunningTime="2025-12-04 18:01:59.899326465 +0000 UTC m=+1381.854687531" watchObservedRunningTime="2025-12-04 18:01:59.908344694 +0000 UTC m=+1381.863705740" Dec 04 18:01:59 crc kubenswrapper[4733]: I1204 18:01:59.918723 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.5987586350000003 podStartE2EDuration="6.91869979s" podCreationTimestamp="2025-12-04 18:01:53 +0000 UTC" firstStartedPulling="2025-12-04 18:01:54.904573912 +0000 UTC m=+1376.859934958" lastFinishedPulling="2025-12-04 18:01:59.224515067 +0000 UTC m=+1381.179876113" observedRunningTime="2025-12-04 18:01:59.91382596 +0000 UTC m=+1381.869187006" watchObservedRunningTime="2025-12-04 18:01:59.91869979 +0000 UTC m=+1381.874060836" Dec 04 18:02:00 crc kubenswrapper[4733]: I1204 18:02:00.867809 4733 generic.go:334] "Generic (PLEG): container finished" podID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerID="32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974" exitCode=143 Dec 04 18:02:00 crc kubenswrapper[4733]: I1204 18:02:00.867901 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0963c49e-2e7a-4ae8-b6d7-55f1926d7476","Type":"ContainerDied","Data":"32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974"} Dec 04 18:02:01 crc kubenswrapper[4733]: I1204 18:02:01.881038 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerStarted","Data":"0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c"} Dec 04 18:02:01 crc kubenswrapper[4733]: I1204 18:02:01.882961 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 18:02:01 crc kubenswrapper[4733]: I1204 18:02:01.920333 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.877749049 podStartE2EDuration="9.920312247s" podCreationTimestamp="2025-12-04 18:01:52 +0000 UTC" firstStartedPulling="2025-12-04 18:01:53.884200349 +0000 UTC m=+1375.839561395" lastFinishedPulling="2025-12-04 18:02:00.926763547 +0000 UTC m=+1382.882124593" observedRunningTime="2025-12-04 18:02:01.904444157 +0000 UTC m=+1383.859805223" watchObservedRunningTime="2025-12-04 18:02:01.920312247 +0000 UTC m=+1383.875673303" Dec 04 18:02:02 crc kubenswrapper[4733]: I1204 18:02:02.904424 4733 generic.go:334] "Generic (PLEG): container finished" podID="553ecb2b-6910-4f64-ac13-93fa1afe5137" containerID="b5cab691f053ad9777f8ae8b5b15035ab043220183fb8752c859b471c466a3d8" exitCode=0 Dec 04 18:02:02 crc kubenswrapper[4733]: I1204 18:02:02.904501 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xv4z2" event={"ID":"553ecb2b-6910-4f64-ac13-93fa1afe5137","Type":"ContainerDied","Data":"b5cab691f053ad9777f8ae8b5b15035ab043220183fb8752c859b471c466a3d8"} Dec 04 18:02:03 crc kubenswrapper[4733]: I1204 18:02:03.925647 4733 generic.go:334] "Generic (PLEG): container finished" podID="18512067-9d64-4796-9bc1-a3d83ef5b13e" containerID="9faff0680efdb357f0f7ab94276aac3340ec7dc913f22f91f8b0cae7b5683fd3" exitCode=0 Dec 04 18:02:03 crc kubenswrapper[4733]: I1204 18:02:03.925734 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" event={"ID":"18512067-9d64-4796-9bc1-a3d83ef5b13e","Type":"ContainerDied","Data":"9faff0680efdb357f0f7ab94276aac3340ec7dc913f22f91f8b0cae7b5683fd3"} Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.066067 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.066144 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.166258 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.166543 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.280653 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.280731 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.313510 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.354703 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.368104 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.400657 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.473285 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5czz5\" (UniqueName: \"kubernetes.io/projected/553ecb2b-6910-4f64-ac13-93fa1afe5137-kube-api-access-5czz5\") pod \"553ecb2b-6910-4f64-ac13-93fa1afe5137\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.473353 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-scripts\") pod \"553ecb2b-6910-4f64-ac13-93fa1afe5137\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.473512 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-combined-ca-bundle\") pod \"553ecb2b-6910-4f64-ac13-93fa1afe5137\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.473547 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-config-data\") pod \"553ecb2b-6910-4f64-ac13-93fa1afe5137\" (UID: \"553ecb2b-6910-4f64-ac13-93fa1afe5137\") " Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.501013 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-scripts" (OuterVolumeSpecName: "scripts") pod "553ecb2b-6910-4f64-ac13-93fa1afe5137" (UID: "553ecb2b-6910-4f64-ac13-93fa1afe5137"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.521225 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/553ecb2b-6910-4f64-ac13-93fa1afe5137-kube-api-access-5czz5" (OuterVolumeSpecName: "kube-api-access-5czz5") pod "553ecb2b-6910-4f64-ac13-93fa1afe5137" (UID: "553ecb2b-6910-4f64-ac13-93fa1afe5137"). InnerVolumeSpecName "kube-api-access-5czz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.590379 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "553ecb2b-6910-4f64-ac13-93fa1afe5137" (UID: "553ecb2b-6910-4f64-ac13-93fa1afe5137"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.601180 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.601244 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5czz5\" (UniqueName: \"kubernetes.io/projected/553ecb2b-6910-4f64-ac13-93fa1afe5137-kube-api-access-5czz5\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.601254 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.608920 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-config-data" (OuterVolumeSpecName: "config-data") pod "553ecb2b-6910-4f64-ac13-93fa1afe5137" (UID: "553ecb2b-6910-4f64-ac13-93fa1afe5137"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.616947 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-922zq"] Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.617181 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" podUID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" containerName="dnsmasq-dns" containerID="cri-o://b83a809aa02d2009fa0815c6ef37a6f8e9afc105fbbb1fbd2aaba3ca86931a93" gracePeriod=10 Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.703231 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/553ecb2b-6910-4f64-ac13-93fa1afe5137-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.950584 4733 generic.go:334] "Generic (PLEG): container finished" podID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" containerID="b83a809aa02d2009fa0815c6ef37a6f8e9afc105fbbb1fbd2aaba3ca86931a93" exitCode=0 Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.950657 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" event={"ID":"dffa9d54-372c-4a62-8955-f0ae3ae2f89a","Type":"ContainerDied","Data":"b83a809aa02d2009fa0815c6ef37a6f8e9afc105fbbb1fbd2aaba3ca86931a93"} Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.953233 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xv4z2" event={"ID":"553ecb2b-6910-4f64-ac13-93fa1afe5137","Type":"ContainerDied","Data":"278ce4720c576463280f17cef3ba35d8061f00ea087467269cfbe9137c96787d"} Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.953353 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="278ce4720c576463280f17cef3ba35d8061f00ea087467269cfbe9137c96787d" Dec 04 18:02:04 crc kubenswrapper[4733]: I1204 18:02:04.953617 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xv4z2" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.003462 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.098449 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.120364 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.120570 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-log" containerID="cri-o://a663e1bc5727c5ca185760cb484948770d7432190c4cda6e4378bca5a497f3b1" gracePeriod=30 Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.121685 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-api" containerID="cri-o://6602dc5879168dcaaa99620ea72395421ca373d0f88940c5eada9244c55900a0" gracePeriod=30 Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.132049 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.132345 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.213503 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-svc\") pod \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.213594 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-sb\") pod \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.213712 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv9x2\" (UniqueName: \"kubernetes.io/projected/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-kube-api-access-lv9x2\") pod \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.213745 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-swift-storage-0\") pod \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.213771 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-config\") pod \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.213798 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-nb\") pod \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\" (UID: \"dffa9d54-372c-4a62-8955-f0ae3ae2f89a\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.227352 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-kube-api-access-lv9x2" (OuterVolumeSpecName: "kube-api-access-lv9x2") pod "dffa9d54-372c-4a62-8955-f0ae3ae2f89a" (UID: "dffa9d54-372c-4a62-8955-f0ae3ae2f89a"). InnerVolumeSpecName "kube-api-access-lv9x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.289217 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dffa9d54-372c-4a62-8955-f0ae3ae2f89a" (UID: "dffa9d54-372c-4a62-8955-f0ae3ae2f89a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.292249 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-config" (OuterVolumeSpecName: "config") pod "dffa9d54-372c-4a62-8955-f0ae3ae2f89a" (UID: "dffa9d54-372c-4a62-8955-f0ae3ae2f89a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.293402 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dffa9d54-372c-4a62-8955-f0ae3ae2f89a" (UID: "dffa9d54-372c-4a62-8955-f0ae3ae2f89a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.308057 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dffa9d54-372c-4a62-8955-f0ae3ae2f89a" (UID: "dffa9d54-372c-4a62-8955-f0ae3ae2f89a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.316305 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.316443 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.316453 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv9x2\" (UniqueName: \"kubernetes.io/projected/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-kube-api-access-lv9x2\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.316461 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.316471 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.325121 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dffa9d54-372c-4a62-8955-f0ae3ae2f89a" (UID: "dffa9d54-372c-4a62-8955-f0ae3ae2f89a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.417718 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dffa9d54-372c-4a62-8955-f0ae3ae2f89a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.422100 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.517370 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.519129 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-scripts\") pod \"18512067-9d64-4796-9bc1-a3d83ef5b13e\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.519756 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-config-data\") pod \"18512067-9d64-4796-9bc1-a3d83ef5b13e\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.519816 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-combined-ca-bundle\") pod \"18512067-9d64-4796-9bc1-a3d83ef5b13e\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.519897 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4h4b\" (UniqueName: \"kubernetes.io/projected/18512067-9d64-4796-9bc1-a3d83ef5b13e-kube-api-access-v4h4b\") pod \"18512067-9d64-4796-9bc1-a3d83ef5b13e\" (UID: \"18512067-9d64-4796-9bc1-a3d83ef5b13e\") " Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.523389 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18512067-9d64-4796-9bc1-a3d83ef5b13e-kube-api-access-v4h4b" (OuterVolumeSpecName: "kube-api-access-v4h4b") pod "18512067-9d64-4796-9bc1-a3d83ef5b13e" (UID: "18512067-9d64-4796-9bc1-a3d83ef5b13e"). InnerVolumeSpecName "kube-api-access-v4h4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.531942 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-scripts" (OuterVolumeSpecName: "scripts") pod "18512067-9d64-4796-9bc1-a3d83ef5b13e" (UID: "18512067-9d64-4796-9bc1-a3d83ef5b13e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.552274 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-config-data" (OuterVolumeSpecName: "config-data") pod "18512067-9d64-4796-9bc1-a3d83ef5b13e" (UID: "18512067-9d64-4796-9bc1-a3d83ef5b13e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.552815 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18512067-9d64-4796-9bc1-a3d83ef5b13e" (UID: "18512067-9d64-4796-9bc1-a3d83ef5b13e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.622446 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.622479 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.622493 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18512067-9d64-4796-9bc1-a3d83ef5b13e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.622506 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4h4b\" (UniqueName: \"kubernetes.io/projected/18512067-9d64-4796-9bc1-a3d83ef5b13e-kube-api-access-v4h4b\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.964387 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerID="a663e1bc5727c5ca185760cb484948770d7432190c4cda6e4378bca5a497f3b1" exitCode=143 Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.964459 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d","Type":"ContainerDied","Data":"a663e1bc5727c5ca185760cb484948770d7432190c4cda6e4378bca5a497f3b1"} Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.967821 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" event={"ID":"18512067-9d64-4796-9bc1-a3d83ef5b13e","Type":"ContainerDied","Data":"9572c9403d011501520ebd9ff21553a210af80301974d5f916fc484ed7bd7348"} Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.967857 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9572c9403d011501520ebd9ff21553a210af80301974d5f916fc484ed7bd7348" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.967949 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8fxb9" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.979002 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.984731 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8c9c7-922zq" event={"ID":"dffa9d54-372c-4a62-8955-f0ae3ae2f89a","Type":"ContainerDied","Data":"7aef1254ec57ad7dc144e5aa1863c799c9a9e0c0376da4ce9612ed438d7308b3"} Dec 04 18:02:05 crc kubenswrapper[4733]: I1204 18:02:05.984810 4733 scope.go:117] "RemoveContainer" containerID="b83a809aa02d2009fa0815c6ef37a6f8e9afc105fbbb1fbd2aaba3ca86931a93" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.034785 4733 scope.go:117] "RemoveContainer" containerID="138f9c8aa9a6459369e3d5b34cfe94fe9cdb0f0de5f46db942505f32795f02cf" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.045124 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 18:02:06 crc kubenswrapper[4733]: E1204 18:02:06.046116 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" containerName="init" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.046266 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" containerName="init" Dec 04 18:02:06 crc kubenswrapper[4733]: E1204 18:02:06.046400 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" containerName="dnsmasq-dns" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.046529 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" containerName="dnsmasq-dns" Dec 04 18:02:06 crc kubenswrapper[4733]: E1204 18:02:06.046628 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553ecb2b-6910-4f64-ac13-93fa1afe5137" containerName="nova-manage" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.046709 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="553ecb2b-6910-4f64-ac13-93fa1afe5137" containerName="nova-manage" Dec 04 18:02:06 crc kubenswrapper[4733]: E1204 18:02:06.046784 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18512067-9d64-4796-9bc1-a3d83ef5b13e" containerName="nova-cell1-conductor-db-sync" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.046891 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="18512067-9d64-4796-9bc1-a3d83ef5b13e" containerName="nova-cell1-conductor-db-sync" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.047292 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="18512067-9d64-4796-9bc1-a3d83ef5b13e" containerName="nova-cell1-conductor-db-sync" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.047383 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="553ecb2b-6910-4f64-ac13-93fa1afe5137" containerName="nova-manage" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.047456 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" containerName="dnsmasq-dns" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.058076 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.059053 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-922zq"] Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.060481 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.072334 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77d8c9c7-922zq"] Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.094852 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.235117 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.235164 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4z5m\" (UniqueName: \"kubernetes.io/projected/839d7537-ff12-4ee6-b36c-b192f130a6e4-kube-api-access-q4z5m\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.235256 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.339193 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.339631 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4z5m\" (UniqueName: \"kubernetes.io/projected/839d7537-ff12-4ee6-b36c-b192f130a6e4-kube-api-access-q4z5m\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.339790 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.345508 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.348107 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.350304 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dffa9d54-372c-4a62-8955-f0ae3ae2f89a" path="/var/lib/kubelet/pods/dffa9d54-372c-4a62-8955-f0ae3ae2f89a/volumes" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.367235 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4z5m\" (UniqueName: \"kubernetes.io/projected/839d7537-ff12-4ee6-b36c-b192f130a6e4-kube-api-access-q4z5m\") pod \"nova-cell1-conductor-0\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.391858 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:06 crc kubenswrapper[4733]: I1204 18:02:06.923615 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 18:02:06 crc kubenswrapper[4733]: W1204 18:02:06.931943 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod839d7537_ff12_4ee6_b36c_b192f130a6e4.slice/crio-8ad19b34b1b777b6642864ee77eac4ec85d9d00ac31f2123a546d6df01a8a97b WatchSource:0}: Error finding container 8ad19b34b1b777b6642864ee77eac4ec85d9d00ac31f2123a546d6df01a8a97b: Status 404 returned error can't find the container with id 8ad19b34b1b777b6642864ee77eac4ec85d9d00ac31f2123a546d6df01a8a97b Dec 04 18:02:07 crc kubenswrapper[4733]: I1204 18:02:07.004471 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"839d7537-ff12-4ee6-b36c-b192f130a6e4","Type":"ContainerStarted","Data":"8ad19b34b1b777b6642864ee77eac4ec85d9d00ac31f2123a546d6df01a8a97b"} Dec 04 18:02:07 crc kubenswrapper[4733]: I1204 18:02:07.004476 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f01545db-3507-4b75-ba11-b37afd1770d7" containerName="nova-scheduler-scheduler" containerID="cri-o://f9319c0a859fc9df0172277a65be92b8bdf0941cf298061e0067aeb25c1626f3" gracePeriod=30 Dec 04 18:02:08 crc kubenswrapper[4733]: I1204 18:02:08.020246 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"839d7537-ff12-4ee6-b36c-b192f130a6e4","Type":"ContainerStarted","Data":"bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f"} Dec 04 18:02:08 crc kubenswrapper[4733]: I1204 18:02:08.020582 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:08 crc kubenswrapper[4733]: I1204 18:02:08.050693 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.050664595 podStartE2EDuration="2.050664595s" podCreationTimestamp="2025-12-04 18:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:08.038121101 +0000 UTC m=+1389.993482177" watchObservedRunningTime="2025-12-04 18:02:08.050664595 +0000 UTC m=+1390.006025661" Dec 04 18:02:09 crc kubenswrapper[4733]: E1204 18:02:09.282051 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f9319c0a859fc9df0172277a65be92b8bdf0941cf298061e0067aeb25c1626f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:02:09 crc kubenswrapper[4733]: E1204 18:02:09.283428 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f9319c0a859fc9df0172277a65be92b8bdf0941cf298061e0067aeb25c1626f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:02:09 crc kubenswrapper[4733]: E1204 18:02:09.284430 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f9319c0a859fc9df0172277a65be92b8bdf0941cf298061e0067aeb25c1626f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:02:09 crc kubenswrapper[4733]: E1204 18:02:09.284464 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f01545db-3507-4b75-ba11-b37afd1770d7" containerName="nova-scheduler-scheduler" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.066831 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d","Type":"ContainerDied","Data":"6602dc5879168dcaaa99620ea72395421ca373d0f88940c5eada9244c55900a0"} Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.066771 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerID="6602dc5879168dcaaa99620ea72395421ca373d0f88940c5eada9244c55900a0" exitCode=0 Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.067170 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d","Type":"ContainerDied","Data":"5adb5f7f07512cf6b3b88cf007522a9b618d4c80fdf3d8b97d98eb11296215c4"} Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.067185 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5adb5f7f07512cf6b3b88cf007522a9b618d4c80fdf3d8b97d98eb11296215c4" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.070074 4733 generic.go:334] "Generic (PLEG): container finished" podID="f01545db-3507-4b75-ba11-b37afd1770d7" containerID="f9319c0a859fc9df0172277a65be92b8bdf0941cf298061e0067aeb25c1626f3" exitCode=0 Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.070103 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f01545db-3507-4b75-ba11-b37afd1770d7","Type":"ContainerDied","Data":"f9319c0a859fc9df0172277a65be92b8bdf0941cf298061e0067aeb25c1626f3"} Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.070548 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.233124 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.251455 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-config-data\") pod \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.251519 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-combined-ca-bundle\") pod \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.251573 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw6zx\" (UniqueName: \"kubernetes.io/projected/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-kube-api-access-gw6zx\") pod \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.251760 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-logs\") pod \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\" (UID: \"5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d\") " Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.252653 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-logs" (OuterVolumeSpecName: "logs") pod "5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" (UID: "5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.263679 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-kube-api-access-gw6zx" (OuterVolumeSpecName: "kube-api-access-gw6zx") pod "5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" (UID: "5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d"). InnerVolumeSpecName "kube-api-access-gw6zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.292125 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" (UID: "5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.300631 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-config-data" (OuterVolumeSpecName: "config-data") pod "5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" (UID: "5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.354749 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-config-data\") pod \"f01545db-3507-4b75-ba11-b37afd1770d7\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.354844 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-combined-ca-bundle\") pod \"f01545db-3507-4b75-ba11-b37afd1770d7\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.354893 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6bgq\" (UniqueName: \"kubernetes.io/projected/f01545db-3507-4b75-ba11-b37afd1770d7-kube-api-access-l6bgq\") pod \"f01545db-3507-4b75-ba11-b37afd1770d7\" (UID: \"f01545db-3507-4b75-ba11-b37afd1770d7\") " Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.355298 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.355316 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.355324 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.355334 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw6zx\" (UniqueName: \"kubernetes.io/projected/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d-kube-api-access-gw6zx\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.359026 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f01545db-3507-4b75-ba11-b37afd1770d7-kube-api-access-l6bgq" (OuterVolumeSpecName: "kube-api-access-l6bgq") pod "f01545db-3507-4b75-ba11-b37afd1770d7" (UID: "f01545db-3507-4b75-ba11-b37afd1770d7"). InnerVolumeSpecName "kube-api-access-l6bgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.378166 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f01545db-3507-4b75-ba11-b37afd1770d7" (UID: "f01545db-3507-4b75-ba11-b37afd1770d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.382021 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-config-data" (OuterVolumeSpecName: "config-data") pod "f01545db-3507-4b75-ba11-b37afd1770d7" (UID: "f01545db-3507-4b75-ba11-b37afd1770d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.457502 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.457538 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01545db-3507-4b75-ba11-b37afd1770d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:11 crc kubenswrapper[4733]: I1204 18:02:11.457552 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6bgq\" (UniqueName: \"kubernetes.io/projected/f01545db-3507-4b75-ba11-b37afd1770d7-kube-api-access-l6bgq\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.081516 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.081534 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.081571 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f01545db-3507-4b75-ba11-b37afd1770d7","Type":"ContainerDied","Data":"5af90d347b9c133c79931454bf8438a50241b223d51c9c7a5a04f5de2f51219c"} Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.083503 4733 scope.go:117] "RemoveContainer" containerID="f9319c0a859fc9df0172277a65be92b8bdf0941cf298061e0067aeb25c1626f3" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.123257 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.135435 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.147963 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.157568 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.166846 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: E1204 18:02:12.167211 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01545db-3507-4b75-ba11-b37afd1770d7" containerName="nova-scheduler-scheduler" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.167222 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01545db-3507-4b75-ba11-b37afd1770d7" containerName="nova-scheduler-scheduler" Dec 04 18:02:12 crc kubenswrapper[4733]: E1204 18:02:12.167234 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-log" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.167240 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-log" Dec 04 18:02:12 crc kubenswrapper[4733]: E1204 18:02:12.167253 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-api" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.167259 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-api" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.167458 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f01545db-3507-4b75-ba11-b37afd1770d7" containerName="nova-scheduler-scheduler" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.167471 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-log" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.167482 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" containerName="nova-api-api" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.168406 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.171913 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.175358 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.187504 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.188965 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.195066 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.212652 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.272250 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h69n\" (UniqueName: \"kubernetes.io/projected/444f1c84-6911-4a47-8435-a1a2369c9766-kube-api-access-8h69n\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.272320 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-config-data\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.272348 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.272392 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/444f1c84-6911-4a47-8435-a1a2369c9766-logs\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.345341 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d" path="/var/lib/kubelet/pods/5a98ae96-6a22-4fcb-8bb5-e7bd45f78b6d/volumes" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.345965 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f01545db-3507-4b75-ba11-b37afd1770d7" path="/var/lib/kubelet/pods/f01545db-3507-4b75-ba11-b37afd1770d7/volumes" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.373951 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h69n\" (UniqueName: \"kubernetes.io/projected/444f1c84-6911-4a47-8435-a1a2369c9766-kube-api-access-8h69n\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.374001 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-config-data\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.374030 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-config-data\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.374052 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.374089 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/444f1c84-6911-4a47-8435-a1a2369c9766-logs\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.374147 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.374202 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz8mc\" (UniqueName: \"kubernetes.io/projected/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-kube-api-access-jz8mc\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.375042 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/444f1c84-6911-4a47-8435-a1a2369c9766-logs\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.379249 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-config-data\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.379394 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.391324 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h69n\" (UniqueName: \"kubernetes.io/projected/444f1c84-6911-4a47-8435-a1a2369c9766-kube-api-access-8h69n\") pod \"nova-api-0\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.476058 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz8mc\" (UniqueName: \"kubernetes.io/projected/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-kube-api-access-jz8mc\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.476201 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-config-data\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.476326 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.482535 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-config-data\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.482607 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.487738 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.497015 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz8mc\" (UniqueName: \"kubernetes.io/projected/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-kube-api-access-jz8mc\") pod \"nova-scheduler-0\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.509997 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:02:12 crc kubenswrapper[4733]: I1204 18:02:12.964626 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:12 crc kubenswrapper[4733]: W1204 18:02:12.966007 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod444f1c84_6911_4a47_8435_a1a2369c9766.slice/crio-cf17f9bf2fc92a3f372cd3e101bd2af9c2c570d490a85965f7dd05b7df966af6 WatchSource:0}: Error finding container cf17f9bf2fc92a3f372cd3e101bd2af9c2c570d490a85965f7dd05b7df966af6: Status 404 returned error can't find the container with id cf17f9bf2fc92a3f372cd3e101bd2af9c2c570d490a85965f7dd05b7df966af6 Dec 04 18:02:13 crc kubenswrapper[4733]: I1204 18:02:13.017444 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:13 crc kubenswrapper[4733]: W1204 18:02:13.018941 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb3677c9_1c3f_4e2d_80d4_64fbe83fead2.slice/crio-9aee316e3e8f29a650a0abf47d49047f12930001f57b3a25aabf7d5ea434b90f WatchSource:0}: Error finding container 9aee316e3e8f29a650a0abf47d49047f12930001f57b3a25aabf7d5ea434b90f: Status 404 returned error can't find the container with id 9aee316e3e8f29a650a0abf47d49047f12930001f57b3a25aabf7d5ea434b90f Dec 04 18:02:13 crc kubenswrapper[4733]: I1204 18:02:13.095181 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2","Type":"ContainerStarted","Data":"9aee316e3e8f29a650a0abf47d49047f12930001f57b3a25aabf7d5ea434b90f"} Dec 04 18:02:13 crc kubenswrapper[4733]: I1204 18:02:13.096475 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"444f1c84-6911-4a47-8435-a1a2369c9766","Type":"ContainerStarted","Data":"cf17f9bf2fc92a3f372cd3e101bd2af9c2c570d490a85965f7dd05b7df966af6"} Dec 04 18:02:14 crc kubenswrapper[4733]: I1204 18:02:14.110979 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2","Type":"ContainerStarted","Data":"756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d"} Dec 04 18:02:14 crc kubenswrapper[4733]: I1204 18:02:14.119613 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"444f1c84-6911-4a47-8435-a1a2369c9766","Type":"ContainerStarted","Data":"bf5ebee771618e639d54ba3ab7b39f700404d7c799700d4ff54261192b7222ef"} Dec 04 18:02:14 crc kubenswrapper[4733]: I1204 18:02:14.119680 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"444f1c84-6911-4a47-8435-a1a2369c9766","Type":"ContainerStarted","Data":"a6a3a63a6eedbc214a7b0744ab707577cd4c5918535ef81fb2827babd93b627e"} Dec 04 18:02:14 crc kubenswrapper[4733]: I1204 18:02:14.130694 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.130672194 podStartE2EDuration="2.130672194s" podCreationTimestamp="2025-12-04 18:02:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:14.129969815 +0000 UTC m=+1396.085330901" watchObservedRunningTime="2025-12-04 18:02:14.130672194 +0000 UTC m=+1396.086033270" Dec 04 18:02:14 crc kubenswrapper[4733]: I1204 18:02:14.169905 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.169879605 podStartE2EDuration="2.169879605s" podCreationTimestamp="2025-12-04 18:02:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:14.158589935 +0000 UTC m=+1396.113951021" watchObservedRunningTime="2025-12-04 18:02:14.169879605 +0000 UTC m=+1396.125240691" Dec 04 18:02:15 crc kubenswrapper[4733]: I1204 18:02:15.361596 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:02:15 crc kubenswrapper[4733]: I1204 18:02:15.361993 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:02:16 crc kubenswrapper[4733]: I1204 18:02:16.441532 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 18:02:17 crc kubenswrapper[4733]: I1204 18:02:17.510200 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 18:02:22 crc kubenswrapper[4733]: I1204 18:02:22.488290 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 18:02:22 crc kubenswrapper[4733]: I1204 18:02:22.488966 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 18:02:22 crc kubenswrapper[4733]: I1204 18:02:22.510466 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 18:02:22 crc kubenswrapper[4733]: I1204 18:02:22.561891 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 18:02:23 crc kubenswrapper[4733]: I1204 18:02:23.243185 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 18:02:23 crc kubenswrapper[4733]: I1204 18:02:23.375971 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 18:02:23 crc kubenswrapper[4733]: I1204 18:02:23.571183 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 18:02:23 crc kubenswrapper[4733]: I1204 18:02:23.571239 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 18:02:27 crc kubenswrapper[4733]: I1204 18:02:27.032273 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:02:27 crc kubenswrapper[4733]: I1204 18:02:27.033045 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="47e99629-3bc7-4756-9b65-11ab755f49ad" containerName="kube-state-metrics" containerID="cri-o://29a65e283a9d82be93f99207a11c39d986879b38e8854120db8b1b0e54553298" gracePeriod=30 Dec 04 18:02:27 crc kubenswrapper[4733]: I1204 18:02:27.271494 4733 generic.go:334] "Generic (PLEG): container finished" podID="47e99629-3bc7-4756-9b65-11ab755f49ad" containerID="29a65e283a9d82be93f99207a11c39d986879b38e8854120db8b1b0e54553298" exitCode=2 Dec 04 18:02:27 crc kubenswrapper[4733]: I1204 18:02:27.271534 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"47e99629-3bc7-4756-9b65-11ab755f49ad","Type":"ContainerDied","Data":"29a65e283a9d82be93f99207a11c39d986879b38e8854120db8b1b0e54553298"} Dec 04 18:02:27 crc kubenswrapper[4733]: I1204 18:02:27.568989 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 18:02:27 crc kubenswrapper[4733]: I1204 18:02:27.589876 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbsgv\" (UniqueName: \"kubernetes.io/projected/47e99629-3bc7-4756-9b65-11ab755f49ad-kube-api-access-fbsgv\") pod \"47e99629-3bc7-4756-9b65-11ab755f49ad\" (UID: \"47e99629-3bc7-4756-9b65-11ab755f49ad\") " Dec 04 18:02:27 crc kubenswrapper[4733]: I1204 18:02:27.620201 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47e99629-3bc7-4756-9b65-11ab755f49ad-kube-api-access-fbsgv" (OuterVolumeSpecName: "kube-api-access-fbsgv") pod "47e99629-3bc7-4756-9b65-11ab755f49ad" (UID: "47e99629-3bc7-4756-9b65-11ab755f49ad"). InnerVolumeSpecName "kube-api-access-fbsgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:27 crc kubenswrapper[4733]: I1204 18:02:27.691654 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbsgv\" (UniqueName: \"kubernetes.io/projected/47e99629-3bc7-4756-9b65-11ab755f49ad-kube-api-access-fbsgv\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.280358 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"47e99629-3bc7-4756-9b65-11ab755f49ad","Type":"ContainerDied","Data":"680494eef17cf2b64057b1af3d5744b0d26f565905768805633001a6fbecc5b9"} Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.280413 4733 scope.go:117] "RemoveContainer" containerID="29a65e283a9d82be93f99207a11c39d986879b38e8854120db8b1b0e54553298" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.280740 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.318286 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.331952 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.347492 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47e99629-3bc7-4756-9b65-11ab755f49ad" path="/var/lib/kubelet/pods/47e99629-3bc7-4756-9b65-11ab755f49ad/volumes" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.348235 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:02:28 crc kubenswrapper[4733]: E1204 18:02:28.348506 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e99629-3bc7-4756-9b65-11ab755f49ad" containerName="kube-state-metrics" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.348520 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e99629-3bc7-4756-9b65-11ab755f49ad" containerName="kube-state-metrics" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.348704 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e99629-3bc7-4756-9b65-11ab755f49ad" containerName="kube-state-metrics" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.349305 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.356453 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.358189 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.358451 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.401681 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.401752 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.401836 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.401870 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjmk2\" (UniqueName: \"kubernetes.io/projected/658fbbbc-cb7d-4795-be06-76a46b589943-kube-api-access-sjmk2\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.503328 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjmk2\" (UniqueName: \"kubernetes.io/projected/658fbbbc-cb7d-4795-be06-76a46b589943-kube-api-access-sjmk2\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.503649 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.503855 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.504063 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.510168 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.511069 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.518641 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjmk2\" (UniqueName: \"kubernetes.io/projected/658fbbbc-cb7d-4795-be06-76a46b589943-kube-api-access-sjmk2\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.526626 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " pod="openstack/kube-state-metrics-0" Dec 04 18:02:28 crc kubenswrapper[4733]: I1204 18:02:28.672232 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.007175 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.008682 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="ceilometer-central-agent" containerID="cri-o://3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112" gracePeriod=30 Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.008845 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="proxy-httpd" containerID="cri-o://0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c" gracePeriod=30 Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.008884 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="sg-core" containerID="cri-o://82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080" gracePeriod=30 Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.008914 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="ceilometer-notification-agent" containerID="cri-o://f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589" gracePeriod=30 Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.181888 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:02:29 crc kubenswrapper[4733]: W1204 18:02:29.185982 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod658fbbbc_cb7d_4795_be06_76a46b589943.slice/crio-1fd5839cf44a213dade16102c343e15d5eebb9a683216736c05d552a656294f6 WatchSource:0}: Error finding container 1fd5839cf44a213dade16102c343e15d5eebb9a683216736c05d552a656294f6: Status 404 returned error can't find the container with id 1fd5839cf44a213dade16102c343e15d5eebb9a683216736c05d552a656294f6 Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.189344 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.291774 4733 generic.go:334] "Generic (PLEG): container finished" podID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerID="0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c" exitCode=0 Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.291826 4733 generic.go:334] "Generic (PLEG): container finished" podID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerID="82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080" exitCode=2 Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.291838 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerDied","Data":"0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c"} Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.291881 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerDied","Data":"82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080"} Dec 04 18:02:29 crc kubenswrapper[4733]: I1204 18:02:29.294168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"658fbbbc-cb7d-4795-be06-76a46b589943","Type":"ContainerStarted","Data":"1fd5839cf44a213dade16102c343e15d5eebb9a683216736c05d552a656294f6"} Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.228355 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.285306 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.309719 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"658fbbbc-cb7d-4795-be06-76a46b589943","Type":"ContainerStarted","Data":"bd700eb5d3983184b0c921d502c52f9cada1865ba48dea23580767a8dbb28917"} Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.310990 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.312917 4733 generic.go:334] "Generic (PLEG): container finished" podID="76ab1aa1-894a-49b4-a1be-8b91429bf2ed" containerID="47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3" exitCode=137 Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.312940 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.312953 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"76ab1aa1-894a-49b4-a1be-8b91429bf2ed","Type":"ContainerDied","Data":"47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3"} Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.313543 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"76ab1aa1-894a-49b4-a1be-8b91429bf2ed","Type":"ContainerDied","Data":"680b7bc2eebd0baa9426d98543482e1111073c5cb7b2f27cd2b0fba415bf199a"} Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.313626 4733 scope.go:117] "RemoveContainer" containerID="47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.317418 4733 generic.go:334] "Generic (PLEG): container finished" podID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerID="5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2" exitCode=137 Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.317466 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.317515 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0963c49e-2e7a-4ae8-b6d7-55f1926d7476","Type":"ContainerDied","Data":"5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2"} Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.317538 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0963c49e-2e7a-4ae8-b6d7-55f1926d7476","Type":"ContainerDied","Data":"662c9edeb67568f3f2b5b3797cb11b0e3e318d16977d0958dc8fec6588c69474"} Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.320250 4733 generic.go:334] "Generic (PLEG): container finished" podID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerID="3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112" exitCode=0 Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.320318 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerDied","Data":"3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112"} Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.337930 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjlp4\" (UniqueName: \"kubernetes.io/projected/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-kube-api-access-mjlp4\") pod \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.337986 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-combined-ca-bundle\") pod \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.338014 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc92j\" (UniqueName: \"kubernetes.io/projected/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-kube-api-access-kc92j\") pod \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.338067 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-combined-ca-bundle\") pod \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.338098 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-config-data\") pod \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\" (UID: \"76ab1aa1-894a-49b4-a1be-8b91429bf2ed\") " Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.338134 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-logs\") pod \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.338176 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-config-data\") pod \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\" (UID: \"0963c49e-2e7a-4ae8-b6d7-55f1926d7476\") " Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.339217 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-logs" (OuterVolumeSpecName: "logs") pod "0963c49e-2e7a-4ae8-b6d7-55f1926d7476" (UID: "0963c49e-2e7a-4ae8-b6d7-55f1926d7476"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.340742 4733 scope.go:117] "RemoveContainer" containerID="47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3" Dec 04 18:02:30 crc kubenswrapper[4733]: E1204 18:02:30.341484 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3\": container with ID starting with 47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3 not found: ID does not exist" containerID="47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.341517 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3"} err="failed to get container status \"47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3\": rpc error: code = NotFound desc = could not find container \"47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3\": container with ID starting with 47733a6040ed23cbadef09ae3b1031d15f41c6c300e354c9ed1196c05d00d4b3 not found: ID does not exist" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.341537 4733 scope.go:117] "RemoveContainer" containerID="5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.353064 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-kube-api-access-kc92j" (OuterVolumeSpecName: "kube-api-access-kc92j") pod "0963c49e-2e7a-4ae8-b6d7-55f1926d7476" (UID: "0963c49e-2e7a-4ae8-b6d7-55f1926d7476"). InnerVolumeSpecName "kube-api-access-kc92j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.353164 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-kube-api-access-mjlp4" (OuterVolumeSpecName: "kube-api-access-mjlp4") pod "76ab1aa1-894a-49b4-a1be-8b91429bf2ed" (UID: "76ab1aa1-894a-49b4-a1be-8b91429bf2ed"). InnerVolumeSpecName "kube-api-access-mjlp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.367355 4733 scope.go:117] "RemoveContainer" containerID="32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.369781 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-config-data" (OuterVolumeSpecName: "config-data") pod "76ab1aa1-894a-49b4-a1be-8b91429bf2ed" (UID: "76ab1aa1-894a-49b4-a1be-8b91429bf2ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.373001 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0963c49e-2e7a-4ae8-b6d7-55f1926d7476" (UID: "0963c49e-2e7a-4ae8-b6d7-55f1926d7476"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.382831 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-config-data" (OuterVolumeSpecName: "config-data") pod "0963c49e-2e7a-4ae8-b6d7-55f1926d7476" (UID: "0963c49e-2e7a-4ae8-b6d7-55f1926d7476"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.389157 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76ab1aa1-894a-49b4-a1be-8b91429bf2ed" (UID: "76ab1aa1-894a-49b4-a1be-8b91429bf2ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.391526 4733 scope.go:117] "RemoveContainer" containerID="5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2" Dec 04 18:02:30 crc kubenswrapper[4733]: E1204 18:02:30.391900 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2\": container with ID starting with 5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2 not found: ID does not exist" containerID="5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.391926 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2"} err="failed to get container status \"5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2\": rpc error: code = NotFound desc = could not find container \"5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2\": container with ID starting with 5e0c2e09d29c7382a4aa28b1cd55cc01edd1eefbd2a051c3352c794cf9d0a3d2 not found: ID does not exist" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.391945 4733 scope.go:117] "RemoveContainer" containerID="32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974" Dec 04 18:02:30 crc kubenswrapper[4733]: E1204 18:02:30.392265 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974\": container with ID starting with 32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974 not found: ID does not exist" containerID="32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.392284 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974"} err="failed to get container status \"32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974\": rpc error: code = NotFound desc = could not find container \"32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974\": container with ID starting with 32cc3a6edbf82a9819b37471f968192ffe24ffa98dbe04d7f22b0ccc7b01a974 not found: ID does not exist" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.439550 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjlp4\" (UniqueName: \"kubernetes.io/projected/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-kube-api-access-mjlp4\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.439588 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.439598 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc92j\" (UniqueName: \"kubernetes.io/projected/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-kube-api-access-kc92j\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.439606 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.439615 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ab1aa1-894a-49b4-a1be-8b91429bf2ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.439624 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.439633 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0963c49e-2e7a-4ae8-b6d7-55f1926d7476-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.647326 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.286154302 podStartE2EDuration="2.647310772s" podCreationTimestamp="2025-12-04 18:02:28 +0000 UTC" firstStartedPulling="2025-12-04 18:02:29.189085582 +0000 UTC m=+1411.144446628" lastFinishedPulling="2025-12-04 18:02:29.550242052 +0000 UTC m=+1411.505603098" observedRunningTime="2025-12-04 18:02:30.341082371 +0000 UTC m=+1412.296443417" watchObservedRunningTime="2025-12-04 18:02:30.647310772 +0000 UTC m=+1412.602671818" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.654134 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.665509 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.680522 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.689730 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.701878 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:02:30 crc kubenswrapper[4733]: E1204 18:02:30.702295 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerName="nova-metadata-log" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.702315 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerName="nova-metadata-log" Dec 04 18:02:30 crc kubenswrapper[4733]: E1204 18:02:30.702353 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerName="nova-metadata-metadata" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.702364 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerName="nova-metadata-metadata" Dec 04 18:02:30 crc kubenswrapper[4733]: E1204 18:02:30.702378 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ab1aa1-894a-49b4-a1be-8b91429bf2ed" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.702386 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ab1aa1-894a-49b4-a1be-8b91429bf2ed" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.702574 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerName="nova-metadata-metadata" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.702590 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" containerName="nova-metadata-log" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.702609 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ab1aa1-894a-49b4-a1be-8b91429bf2ed" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.703224 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.708878 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.709074 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.709312 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.709541 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.710427 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.712861 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.713043 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.715751 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.724930 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.846824 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lm8w\" (UniqueName: \"kubernetes.io/projected/c1acfe14-408d-4782-a419-9eb454c13de1-kube-api-access-2lm8w\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.846870 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.846920 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-config-data\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.846974 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.846999 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.847021 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.847038 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5868\" (UniqueName: \"kubernetes.io/projected/4047083c-6418-4814-8b24-9407494d0303-kube-api-access-h5868\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.847052 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.847065 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.847083 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1acfe14-408d-4782-a419-9eb454c13de1-logs\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.948137 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lm8w\" (UniqueName: \"kubernetes.io/projected/c1acfe14-408d-4782-a419-9eb454c13de1-kube-api-access-2lm8w\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.948434 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.948584 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-config-data\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.948754 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.948904 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.948997 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.949077 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5868\" (UniqueName: \"kubernetes.io/projected/4047083c-6418-4814-8b24-9407494d0303-kube-api-access-h5868\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.949144 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.949215 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.949334 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1acfe14-408d-4782-a419-9eb454c13de1-logs\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.949909 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1acfe14-408d-4782-a419-9eb454c13de1-logs\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.952534 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.953774 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-config-data\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.954938 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.956727 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.958000 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.958647 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.965670 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.974530 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5868\" (UniqueName: \"kubernetes.io/projected/4047083c-6418-4814-8b24-9407494d0303-kube-api-access-h5868\") pod \"nova-cell1-novncproxy-0\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:30 crc kubenswrapper[4733]: I1204 18:02:30.976083 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lm8w\" (UniqueName: \"kubernetes.io/projected/c1acfe14-408d-4782-a419-9eb454c13de1-kube-api-access-2lm8w\") pod \"nova-metadata-0\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " pod="openstack/nova-metadata-0" Dec 04 18:02:31 crc kubenswrapper[4733]: I1204 18:02:31.027904 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:31 crc kubenswrapper[4733]: I1204 18:02:31.041966 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:02:31 crc kubenswrapper[4733]: I1204 18:02:31.647228 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:02:31 crc kubenswrapper[4733]: I1204 18:02:31.774599 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.346659 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0963c49e-2e7a-4ae8-b6d7-55f1926d7476" path="/var/lib/kubelet/pods/0963c49e-2e7a-4ae8-b6d7-55f1926d7476/volumes" Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.347721 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76ab1aa1-894a-49b4-a1be-8b91429bf2ed" path="/var/lib/kubelet/pods/76ab1aa1-894a-49b4-a1be-8b91429bf2ed/volumes" Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.376647 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4047083c-6418-4814-8b24-9407494d0303","Type":"ContainerStarted","Data":"55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1"} Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.376954 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4047083c-6418-4814-8b24-9407494d0303","Type":"ContainerStarted","Data":"ac2e572ab8366f12b677177b1a49a813aebef2886f0ed56f1e19766e5e6c543d"} Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.379053 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1acfe14-408d-4782-a419-9eb454c13de1","Type":"ContainerStarted","Data":"7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b"} Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.379084 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1acfe14-408d-4782-a419-9eb454c13de1","Type":"ContainerStarted","Data":"070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6"} Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.379096 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1acfe14-408d-4782-a419-9eb454c13de1","Type":"ContainerStarted","Data":"72325e0e45e6cbf2d87a27dcb280f1e775075eec301e8b8c1a4da5cb6e190633"} Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.402699 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.4026754009999998 podStartE2EDuration="2.402675401s" podCreationTimestamp="2025-12-04 18:02:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:32.39133706 +0000 UTC m=+1414.346698126" watchObservedRunningTime="2025-12-04 18:02:32.402675401 +0000 UTC m=+1414.358036457" Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.433213 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.433191372 podStartE2EDuration="2.433191372s" podCreationTimestamp="2025-12-04 18:02:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:32.420132606 +0000 UTC m=+1414.375493662" watchObservedRunningTime="2025-12-04 18:02:32.433191372 +0000 UTC m=+1414.388552418" Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.492844 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.493271 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.496689 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 18:02:32 crc kubenswrapper[4733]: I1204 18:02:32.499430 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.404544 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.416458 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.636885 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-zs7pg"] Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.638353 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.715666 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-zs7pg"] Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.812756 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.813183 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-config\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.813213 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96wk2\" (UniqueName: \"kubernetes.io/projected/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-kube-api-access-96wk2\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.813382 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.813509 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.813578 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.915614 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-config\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.915675 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96wk2\" (UniqueName: \"kubernetes.io/projected/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-kube-api-access-96wk2\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.915758 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.915843 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.915881 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.915939 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.917108 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-config\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.917153 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-sb\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.917096 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-swift-storage-0\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.917386 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-nb\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.918329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-svc\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.927921 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.940715 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96wk2\" (UniqueName: \"kubernetes.io/projected/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-kube-api-access-96wk2\") pod \"dnsmasq-dns-cf4ff87b5-zs7pg\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:33 crc kubenswrapper[4733]: I1204 18:02:33.963358 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.122409 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-scripts\") pod \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.122520 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-run-httpd\") pod \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.122608 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-combined-ca-bundle\") pod \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.122633 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-sg-core-conf-yaml\") pod \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.122722 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-config-data\") pod \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.122778 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-log-httpd\") pod \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.122823 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7zjh\" (UniqueName: \"kubernetes.io/projected/32c62974-76c1-4ce1-8c63-0c231ef8e38c-kube-api-access-f7zjh\") pod \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\" (UID: \"32c62974-76c1-4ce1-8c63-0c231ef8e38c\") " Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.125251 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "32c62974-76c1-4ce1-8c63-0c231ef8e38c" (UID: "32c62974-76c1-4ce1-8c63-0c231ef8e38c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.125408 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "32c62974-76c1-4ce1-8c63-0c231ef8e38c" (UID: "32c62974-76c1-4ce1-8c63-0c231ef8e38c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.128128 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c62974-76c1-4ce1-8c63-0c231ef8e38c-kube-api-access-f7zjh" (OuterVolumeSpecName: "kube-api-access-f7zjh") pod "32c62974-76c1-4ce1-8c63-0c231ef8e38c" (UID: "32c62974-76c1-4ce1-8c63-0c231ef8e38c"). InnerVolumeSpecName "kube-api-access-f7zjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.128991 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-scripts" (OuterVolumeSpecName: "scripts") pod "32c62974-76c1-4ce1-8c63-0c231ef8e38c" (UID: "32c62974-76c1-4ce1-8c63-0c231ef8e38c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.167037 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "32c62974-76c1-4ce1-8c63-0c231ef8e38c" (UID: "32c62974-76c1-4ce1-8c63-0c231ef8e38c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.225986 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.226021 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7zjh\" (UniqueName: \"kubernetes.io/projected/32c62974-76c1-4ce1-8c63-0c231ef8e38c-kube-api-access-f7zjh\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.226033 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.226043 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32c62974-76c1-4ce1-8c63-0c231ef8e38c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.226054 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.234805 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32c62974-76c1-4ce1-8c63-0c231ef8e38c" (UID: "32c62974-76c1-4ce1-8c63-0c231ef8e38c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.253441 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-config-data" (OuterVolumeSpecName: "config-data") pod "32c62974-76c1-4ce1-8c63-0c231ef8e38c" (UID: "32c62974-76c1-4ce1-8c63-0c231ef8e38c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.327670 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.327710 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c62974-76c1-4ce1-8c63-0c231ef8e38c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.424711 4733 generic.go:334] "Generic (PLEG): container finished" podID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerID="f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589" exitCode=0 Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.424774 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerDied","Data":"f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589"} Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.424827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32c62974-76c1-4ce1-8c63-0c231ef8e38c","Type":"ContainerDied","Data":"9197e904a60e2e6b62980c11341a1c87487bcd838cfc655374722c6eb66a56ee"} Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.424837 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.424848 4733 scope.go:117] "RemoveContainer" containerID="0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.455220 4733 scope.go:117] "RemoveContainer" containerID="82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.465067 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.477064 4733 scope.go:117] "RemoveContainer" containerID="f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.505493 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.527295 4733 scope.go:117] "RemoveContainer" containerID="3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112" Dec 04 18:02:34 crc kubenswrapper[4733]: W1204 18:02:34.541055 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7446c8ed_a97b_41a1_ae84_8f1eed8b203b.slice/crio-e7892397ee71ab39db17b4f9ce076b0a9c65abc9b24ca5e76b41e706b70aa675 WatchSource:0}: Error finding container e7892397ee71ab39db17b4f9ce076b0a9c65abc9b24ca5e76b41e706b70aa675: Status 404 returned error can't find the container with id e7892397ee71ab39db17b4f9ce076b0a9c65abc9b24ca5e76b41e706b70aa675 Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.553083 4733 scope.go:117] "RemoveContainer" containerID="0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c" Dec 04 18:02:34 crc kubenswrapper[4733]: E1204 18:02:34.554675 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c\": container with ID starting with 0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c not found: ID does not exist" containerID="0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.554710 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c"} err="failed to get container status \"0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c\": rpc error: code = NotFound desc = could not find container \"0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c\": container with ID starting with 0d28bed253c27eafd289126173a513a4dc30fc4b193aed1eac782768bda85a9c not found: ID does not exist" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.554730 4733 scope.go:117] "RemoveContainer" containerID="82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080" Dec 04 18:02:34 crc kubenswrapper[4733]: E1204 18:02:34.557036 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080\": container with ID starting with 82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080 not found: ID does not exist" containerID="82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.557079 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080"} err="failed to get container status \"82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080\": rpc error: code = NotFound desc = could not find container \"82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080\": container with ID starting with 82da9cd0facf20a15e31cccb2c348733e1c4d5287c6e9f7759a0b15c02ec6080 not found: ID does not exist" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.557106 4733 scope.go:117] "RemoveContainer" containerID="f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.557205 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:34 crc kubenswrapper[4733]: E1204 18:02:34.557509 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589\": container with ID starting with f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589 not found: ID does not exist" containerID="f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.557557 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589"} err="failed to get container status \"f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589\": rpc error: code = NotFound desc = could not find container \"f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589\": container with ID starting with f9fe8c425a1d8a808523b486ed3c4db6565e4724541d715abd6071715207c589 not found: ID does not exist" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.557607 4733 scope.go:117] "RemoveContainer" containerID="3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112" Dec 04 18:02:34 crc kubenswrapper[4733]: E1204 18:02:34.558039 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="sg-core" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558058 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="sg-core" Dec 04 18:02:34 crc kubenswrapper[4733]: E1204 18:02:34.558078 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="ceilometer-notification-agent" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558084 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="ceilometer-notification-agent" Dec 04 18:02:34 crc kubenswrapper[4733]: E1204 18:02:34.558100 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="proxy-httpd" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558106 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="proxy-httpd" Dec 04 18:02:34 crc kubenswrapper[4733]: E1204 18:02:34.558143 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="ceilometer-central-agent" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558149 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="ceilometer-central-agent" Dec 04 18:02:34 crc kubenswrapper[4733]: E1204 18:02:34.558472 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112\": container with ID starting with 3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112 not found: ID does not exist" containerID="3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558519 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112"} err="failed to get container status \"3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112\": rpc error: code = NotFound desc = could not find container \"3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112\": container with ID starting with 3848d259611b2902ffddc8ff2bbf9e974169650decf2bf71d20ada7461fc3112 not found: ID does not exist" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558551 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="proxy-httpd" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558576 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="ceilometer-central-agent" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558594 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="sg-core" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.558612 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" containerName="ceilometer-notification-agent" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.565616 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.567921 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.567960 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.571985 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.573756 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.593618 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-zs7pg"] Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.738701 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-log-httpd\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.739070 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wj7z\" (UniqueName: \"kubernetes.io/projected/d67f59d9-ce48-4306-b966-fdba768dc612-kube-api-access-5wj7z\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.739202 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-scripts\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.739232 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-run-httpd\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.740049 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.740229 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.740287 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.740327 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-config-data\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.842687 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-scripts\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.842757 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-run-httpd\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.842827 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.842858 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.842893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.842909 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-config-data\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.842972 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-log-httpd\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.842992 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wj7z\" (UniqueName: \"kubernetes.io/projected/d67f59d9-ce48-4306-b966-fdba768dc612-kube-api-access-5wj7z\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.843287 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-run-httpd\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.843819 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-log-httpd\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.850623 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.855759 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-config-data\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.856761 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.858574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.859345 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-scripts\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.866363 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wj7z\" (UniqueName: \"kubernetes.io/projected/d67f59d9-ce48-4306-b966-fdba768dc612-kube-api-access-5wj7z\") pod \"ceilometer-0\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " pod="openstack/ceilometer-0" Dec 04 18:02:34 crc kubenswrapper[4733]: I1204 18:02:34.891249 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:02:35 crc kubenswrapper[4733]: I1204 18:02:35.435656 4733 generic.go:334] "Generic (PLEG): container finished" podID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" containerID="40bce759ec7e8d43134a5240b984af64caf430aa9f304d8430db8fd918772895" exitCode=0 Dec 04 18:02:35 crc kubenswrapper[4733]: I1204 18:02:35.435758 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" event={"ID":"7446c8ed-a97b-41a1-ae84-8f1eed8b203b","Type":"ContainerDied","Data":"40bce759ec7e8d43134a5240b984af64caf430aa9f304d8430db8fd918772895"} Dec 04 18:02:35 crc kubenswrapper[4733]: I1204 18:02:35.435977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" event={"ID":"7446c8ed-a97b-41a1-ae84-8f1eed8b203b","Type":"ContainerStarted","Data":"e7892397ee71ab39db17b4f9ce076b0a9c65abc9b24ca5e76b41e706b70aa675"} Dec 04 18:02:35 crc kubenswrapper[4733]: I1204 18:02:35.453651 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:35 crc kubenswrapper[4733]: W1204 18:02:35.458902 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd67f59d9_ce48_4306_b966_fdba768dc612.slice/crio-e4509d4bdeceeb929f056051733e62a9dae0bb7e5fe100aaf0d818bf3badeeb8 WatchSource:0}: Error finding container e4509d4bdeceeb929f056051733e62a9dae0bb7e5fe100aaf0d818bf3badeeb8: Status 404 returned error can't find the container with id e4509d4bdeceeb929f056051733e62a9dae0bb7e5fe100aaf0d818bf3badeeb8 Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.028859 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.043302 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.043353 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.140222 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.345444 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32c62974-76c1-4ce1-8c63-0c231ef8e38c" path="/var/lib/kubelet/pods/32c62974-76c1-4ce1-8c63-0c231ef8e38c/volumes" Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.444877 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerStarted","Data":"1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75"} Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.444935 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerStarted","Data":"e4509d4bdeceeb929f056051733e62a9dae0bb7e5fe100aaf0d818bf3badeeb8"} Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.447225 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" event={"ID":"7446c8ed-a97b-41a1-ae84-8f1eed8b203b","Type":"ContainerStarted","Data":"7ba1c201c7cef13fb8704e43d271cf5b0bf7479dc2ba8022c5b90de90fe6784e"} Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.447428 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:36 crc kubenswrapper[4733]: I1204 18:02:36.470765 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" podStartSLOduration=3.4707397 podStartE2EDuration="3.4707397s" podCreationTimestamp="2025-12-04 18:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:36.464069082 +0000 UTC m=+1418.419430168" watchObservedRunningTime="2025-12-04 18:02:36.4707397 +0000 UTC m=+1418.426100776" Dec 04 18:02:37 crc kubenswrapper[4733]: I1204 18:02:37.462535 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerStarted","Data":"449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882"} Dec 04 18:02:37 crc kubenswrapper[4733]: I1204 18:02:37.708951 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:37 crc kubenswrapper[4733]: I1204 18:02:37.709171 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-log" containerID="cri-o://a6a3a63a6eedbc214a7b0744ab707577cd4c5918535ef81fb2827babd93b627e" gracePeriod=30 Dec 04 18:02:37 crc kubenswrapper[4733]: I1204 18:02:37.709292 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-api" containerID="cri-o://bf5ebee771618e639d54ba3ab7b39f700404d7c799700d4ff54261192b7222ef" gracePeriod=30 Dec 04 18:02:38 crc kubenswrapper[4733]: I1204 18:02:38.473072 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerStarted","Data":"a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2"} Dec 04 18:02:38 crc kubenswrapper[4733]: I1204 18:02:38.475600 4733 generic.go:334] "Generic (PLEG): container finished" podID="444f1c84-6911-4a47-8435-a1a2369c9766" containerID="a6a3a63a6eedbc214a7b0744ab707577cd4c5918535ef81fb2827babd93b627e" exitCode=143 Dec 04 18:02:38 crc kubenswrapper[4733]: I1204 18:02:38.475635 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"444f1c84-6911-4a47-8435-a1a2369c9766","Type":"ContainerDied","Data":"a6a3a63a6eedbc214a7b0744ab707577cd4c5918535ef81fb2827babd93b627e"} Dec 04 18:02:38 crc kubenswrapper[4733]: I1204 18:02:38.684242 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 18:02:39 crc kubenswrapper[4733]: I1204 18:02:39.485951 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerStarted","Data":"bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d"} Dec 04 18:02:39 crc kubenswrapper[4733]: I1204 18:02:39.486083 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="ceilometer-central-agent" containerID="cri-o://1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75" gracePeriod=30 Dec 04 18:02:39 crc kubenswrapper[4733]: I1204 18:02:39.486182 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="ceilometer-notification-agent" containerID="cri-o://449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882" gracePeriod=30 Dec 04 18:02:39 crc kubenswrapper[4733]: I1204 18:02:39.486212 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="proxy-httpd" containerID="cri-o://bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d" gracePeriod=30 Dec 04 18:02:39 crc kubenswrapper[4733]: I1204 18:02:39.486219 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="sg-core" containerID="cri-o://a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2" gracePeriod=30 Dec 04 18:02:39 crc kubenswrapper[4733]: I1204 18:02:39.486538 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 18:02:39 crc kubenswrapper[4733]: I1204 18:02:39.508293 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.072622787 podStartE2EDuration="5.508270882s" podCreationTimestamp="2025-12-04 18:02:34 +0000 UTC" firstStartedPulling="2025-12-04 18:02:35.461662315 +0000 UTC m=+1417.417023361" lastFinishedPulling="2025-12-04 18:02:38.89731041 +0000 UTC m=+1420.852671456" observedRunningTime="2025-12-04 18:02:39.50631247 +0000 UTC m=+1421.461673546" watchObservedRunningTime="2025-12-04 18:02:39.508270882 +0000 UTC m=+1421.463631938" Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.496438 4733 generic.go:334] "Generic (PLEG): container finished" podID="d67f59d9-ce48-4306-b966-fdba768dc612" containerID="bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d" exitCode=0 Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.496813 4733 generic.go:334] "Generic (PLEG): container finished" podID="d67f59d9-ce48-4306-b966-fdba768dc612" containerID="a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2" exitCode=2 Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.496828 4733 generic.go:334] "Generic (PLEG): container finished" podID="d67f59d9-ce48-4306-b966-fdba768dc612" containerID="449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882" exitCode=0 Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.496536 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerDied","Data":"bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d"} Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.496884 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerDied","Data":"a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2"} Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.496904 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerDied","Data":"449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882"} Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.871665 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-thnp8"] Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.873500 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:40 crc kubenswrapper[4733]: I1204 18:02:40.887041 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-thnp8"] Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.028455 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.043038 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.043086 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.049097 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.050059 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-catalog-content\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.050262 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-utilities\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.050303 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvmk9\" (UniqueName: \"kubernetes.io/projected/502fb5bc-ecaa-4a78-92fd-02b8eb138442-kube-api-access-tvmk9\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.152448 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-utilities\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.152538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvmk9\" (UniqueName: \"kubernetes.io/projected/502fb5bc-ecaa-4a78-92fd-02b8eb138442-kube-api-access-tvmk9\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.152699 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-catalog-content\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.153069 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-utilities\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.153221 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-catalog-content\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.178689 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvmk9\" (UniqueName: \"kubernetes.io/projected/502fb5bc-ecaa-4a78-92fd-02b8eb138442-kube-api-access-tvmk9\") pod \"redhat-operators-thnp8\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.194201 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.528344 4733 generic.go:334] "Generic (PLEG): container finished" podID="444f1c84-6911-4a47-8435-a1a2369c9766" containerID="bf5ebee771618e639d54ba3ab7b39f700404d7c799700d4ff54261192b7222ef" exitCode=0 Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.528843 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"444f1c84-6911-4a47-8435-a1a2369c9766","Type":"ContainerDied","Data":"bf5ebee771618e639d54ba3ab7b39f700404d7c799700d4ff54261192b7222ef"} Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.570031 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.604848 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-thnp8"] Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.802013 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-ksgdb"] Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.803400 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.806196 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.806405 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.808545 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ksgdb"] Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.969364 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-config-data\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.969658 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-scripts\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.969736 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:41 crc kubenswrapper[4733]: I1204 18:02:41.969817 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97chn\" (UniqueName: \"kubernetes.io/projected/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-kube-api-access-97chn\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.061547 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.061560 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.071891 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.072009 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97chn\" (UniqueName: \"kubernetes.io/projected/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-kube-api-access-97chn\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.072356 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-config-data\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.072385 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-scripts\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.086534 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-scripts\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.088574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-config-data\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.089523 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.092324 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97chn\" (UniqueName: \"kubernetes.io/projected/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-kube-api-access-97chn\") pod \"nova-cell1-cell-mapping-ksgdb\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.142673 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:42 crc kubenswrapper[4733]: E1204 18:02:42.153989 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod502fb5bc_ecaa_4a78_92fd_02b8eb138442.slice/crio-d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod502fb5bc_ecaa_4a78_92fd_02b8eb138442.slice/crio-conmon-d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5.scope\": RecentStats: unable to find data in memory cache]" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.314031 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.482006 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h69n\" (UniqueName: \"kubernetes.io/projected/444f1c84-6911-4a47-8435-a1a2369c9766-kube-api-access-8h69n\") pod \"444f1c84-6911-4a47-8435-a1a2369c9766\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.482074 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-combined-ca-bundle\") pod \"444f1c84-6911-4a47-8435-a1a2369c9766\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.482100 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-config-data\") pod \"444f1c84-6911-4a47-8435-a1a2369c9766\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.482271 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/444f1c84-6911-4a47-8435-a1a2369c9766-logs\") pod \"444f1c84-6911-4a47-8435-a1a2369c9766\" (UID: \"444f1c84-6911-4a47-8435-a1a2369c9766\") " Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.489176 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/444f1c84-6911-4a47-8435-a1a2369c9766-logs" (OuterVolumeSpecName: "logs") pod "444f1c84-6911-4a47-8435-a1a2369c9766" (UID: "444f1c84-6911-4a47-8435-a1a2369c9766"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.502082 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/444f1c84-6911-4a47-8435-a1a2369c9766-kube-api-access-8h69n" (OuterVolumeSpecName: "kube-api-access-8h69n") pod "444f1c84-6911-4a47-8435-a1a2369c9766" (UID: "444f1c84-6911-4a47-8435-a1a2369c9766"). InnerVolumeSpecName "kube-api-access-8h69n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.533330 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-config-data" (OuterVolumeSpecName: "config-data") pod "444f1c84-6911-4a47-8435-a1a2369c9766" (UID: "444f1c84-6911-4a47-8435-a1a2369c9766"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.542923 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "444f1c84-6911-4a47-8435-a1a2369c9766" (UID: "444f1c84-6911-4a47-8435-a1a2369c9766"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.554217 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.555303 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"444f1c84-6911-4a47-8435-a1a2369c9766","Type":"ContainerDied","Data":"cf17f9bf2fc92a3f372cd3e101bd2af9c2c570d490a85965f7dd05b7df966af6"} Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.555351 4733 scope.go:117] "RemoveContainer" containerID="bf5ebee771618e639d54ba3ab7b39f700404d7c799700d4ff54261192b7222ef" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.565950 4733 generic.go:334] "Generic (PLEG): container finished" podID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerID="d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5" exitCode=0 Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.567938 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thnp8" event={"ID":"502fb5bc-ecaa-4a78-92fd-02b8eb138442","Type":"ContainerDied","Data":"d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5"} Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.567980 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thnp8" event={"ID":"502fb5bc-ecaa-4a78-92fd-02b8eb138442","Type":"ContainerStarted","Data":"a695e2d9962fb2460b3e992875a18ab2f2a5a22562cc2d17ea19b8d38d683b1c"} Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.589490 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h69n\" (UniqueName: \"kubernetes.io/projected/444f1c84-6911-4a47-8435-a1a2369c9766-kube-api-access-8h69n\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.589750 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.589759 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444f1c84-6911-4a47-8435-a1a2369c9766-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.589779 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/444f1c84-6911-4a47-8435-a1a2369c9766-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.627503 4733 scope.go:117] "RemoveContainer" containerID="a6a3a63a6eedbc214a7b0744ab707577cd4c5918535ef81fb2827babd93b627e" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.658039 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.684763 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.727844 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ksgdb"] Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.751412 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:42 crc kubenswrapper[4733]: E1204 18:02:42.752091 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-api" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.752112 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-api" Dec 04 18:02:42 crc kubenswrapper[4733]: E1204 18:02:42.752150 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-log" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.752156 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-log" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.752426 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-log" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.752484 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" containerName="nova-api-api" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.753570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.756896 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.757100 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.757213 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.773768 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.900193 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.900240 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96cwf\" (UniqueName: \"kubernetes.io/projected/c86cb171-a0ab-4a6c-bf23-3b44e981976c-kube-api-access-96cwf\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.900276 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-config-data\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.900423 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.900612 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-public-tls-certs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:42 crc kubenswrapper[4733]: I1204 18:02:42.900641 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86cb171-a0ab-4a6c-bf23-3b44e981976c-logs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.002345 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.002645 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96cwf\" (UniqueName: \"kubernetes.io/projected/c86cb171-a0ab-4a6c-bf23-3b44e981976c-kube-api-access-96cwf\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.002678 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-config-data\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.002741 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.002814 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-public-tls-certs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.002830 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86cb171-a0ab-4a6c-bf23-3b44e981976c-logs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.003132 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86cb171-a0ab-4a6c-bf23-3b44e981976c-logs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.008334 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.008334 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-public-tls-certs\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.008698 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.012193 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-config-data\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.019325 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96cwf\" (UniqueName: \"kubernetes.io/projected/c86cb171-a0ab-4a6c-bf23-3b44e981976c-kube-api-access-96cwf\") pod \"nova-api-0\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.088828 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.275136 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f8brz"] Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.277455 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.283744 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f8brz"] Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.411196 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-utilities\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.411634 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgkjq\" (UniqueName: \"kubernetes.io/projected/8906a11a-4c2c-4e38-865b-664be708e41e-kube-api-access-mgkjq\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.411669 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-catalog-content\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.512980 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgkjq\" (UniqueName: \"kubernetes.io/projected/8906a11a-4c2c-4e38-865b-664be708e41e-kube-api-access-mgkjq\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.513039 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-catalog-content\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.513105 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-utilities\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.513572 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-utilities\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.514080 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-catalog-content\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.532624 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgkjq\" (UniqueName: \"kubernetes.io/projected/8906a11a-4c2c-4e38-865b-664be708e41e-kube-api-access-mgkjq\") pod \"community-operators-f8brz\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.608217 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.608501 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ksgdb" event={"ID":"246f7a9e-8e85-4aa7-8bf4-44739e1d500d","Type":"ContainerStarted","Data":"9a63a5646f24f905ab643629b703d9fa8539f845df76e04dbee266c5247dbf9d"} Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.608554 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ksgdb" event={"ID":"246f7a9e-8e85-4aa7-8bf4-44739e1d500d","Type":"ContainerStarted","Data":"4a0dc5ee94f7b67147c48e4958ce8050e758d62271c8fb681785fcc814368a24"} Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.637665 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-ksgdb" podStartSLOduration=2.637649617 podStartE2EDuration="2.637649617s" podCreationTimestamp="2025-12-04 18:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:43.62909385 +0000 UTC m=+1425.584454896" watchObservedRunningTime="2025-12-04 18:02:43.637649617 +0000 UTC m=+1425.593010663" Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.732819 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:43 crc kubenswrapper[4733]: I1204 18:02:43.966438 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.044886 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-qbzmf"] Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.045141 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" podUID="16c800b6-4095-4676-ac1a-2125b3b40031" containerName="dnsmasq-dns" containerID="cri-o://f32eabfab4f0ea2f0f6ea5778bd2054fd4fbd57466933574641025b23450044f" gracePeriod=10 Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.272149 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f8brz"] Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.391422 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="444f1c84-6911-4a47-8435-a1a2369c9766" path="/var/lib/kubelet/pods/444f1c84-6911-4a47-8435-a1a2369c9766/volumes" Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.402323 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" podUID="16c800b6-4095-4676-ac1a-2125b3b40031" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.190:5353: connect: connection refused" Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.658344 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thnp8" event={"ID":"502fb5bc-ecaa-4a78-92fd-02b8eb138442","Type":"ContainerStarted","Data":"3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac"} Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.687309 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8brz" event={"ID":"8906a11a-4c2c-4e38-865b-664be708e41e","Type":"ContainerStarted","Data":"4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288"} Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.687353 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8brz" event={"ID":"8906a11a-4c2c-4e38-865b-664be708e41e","Type":"ContainerStarted","Data":"1c406b0ada12d9452c7de684f4e7740cd1d530de263ad1bfbc156c9b48f82b41"} Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.700114 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c86cb171-a0ab-4a6c-bf23-3b44e981976c","Type":"ContainerStarted","Data":"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8"} Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.700157 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c86cb171-a0ab-4a6c-bf23-3b44e981976c","Type":"ContainerStarted","Data":"e240ce27cfe5cd8ea2ee788da03ef20acd32be270c77f3e99a57a0f1c2704168"} Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.716051 4733 generic.go:334] "Generic (PLEG): container finished" podID="16c800b6-4095-4676-ac1a-2125b3b40031" containerID="f32eabfab4f0ea2f0f6ea5778bd2054fd4fbd57466933574641025b23450044f" exitCode=0 Dec 04 18:02:44 crc kubenswrapper[4733]: I1204 18:02:44.716235 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" event={"ID":"16c800b6-4095-4676-ac1a-2125b3b40031","Type":"ContainerDied","Data":"f32eabfab4f0ea2f0f6ea5778bd2054fd4fbd57466933574641025b23450044f"} Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.189893 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.287407 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-sb\") pod \"16c800b6-4095-4676-ac1a-2125b3b40031\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.287492 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-config\") pod \"16c800b6-4095-4676-ac1a-2125b3b40031\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.287646 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-swift-storage-0\") pod \"16c800b6-4095-4676-ac1a-2125b3b40031\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.287672 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-nb\") pod \"16c800b6-4095-4676-ac1a-2125b3b40031\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.287879 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-svc\") pod \"16c800b6-4095-4676-ac1a-2125b3b40031\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.287909 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f6bt\" (UniqueName: \"kubernetes.io/projected/16c800b6-4095-4676-ac1a-2125b3b40031-kube-api-access-9f6bt\") pod \"16c800b6-4095-4676-ac1a-2125b3b40031\" (UID: \"16c800b6-4095-4676-ac1a-2125b3b40031\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.296772 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c800b6-4095-4676-ac1a-2125b3b40031-kube-api-access-9f6bt" (OuterVolumeSpecName: "kube-api-access-9f6bt") pod "16c800b6-4095-4676-ac1a-2125b3b40031" (UID: "16c800b6-4095-4676-ac1a-2125b3b40031"). InnerVolumeSpecName "kube-api-access-9f6bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.362340 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.362410 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.362465 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.363324 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1e31ad24991110558bf25e49f4521223e1cb18d3a3ee43198b8eff935ecc2a3"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.363386 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://f1e31ad24991110558bf25e49f4521223e1cb18d3a3ee43198b8eff935ecc2a3" gracePeriod=600 Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.391106 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f6bt\" (UniqueName: \"kubernetes.io/projected/16c800b6-4095-4676-ac1a-2125b3b40031-kube-api-access-9f6bt\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.393304 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "16c800b6-4095-4676-ac1a-2125b3b40031" (UID: "16c800b6-4095-4676-ac1a-2125b3b40031"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.395499 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "16c800b6-4095-4676-ac1a-2125b3b40031" (UID: "16c800b6-4095-4676-ac1a-2125b3b40031"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.415217 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "16c800b6-4095-4676-ac1a-2125b3b40031" (UID: "16c800b6-4095-4676-ac1a-2125b3b40031"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.422110 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-config" (OuterVolumeSpecName: "config") pod "16c800b6-4095-4676-ac1a-2125b3b40031" (UID: "16c800b6-4095-4676-ac1a-2125b3b40031"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.447679 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "16c800b6-4095-4676-ac1a-2125b3b40031" (UID: "16c800b6-4095-4676-ac1a-2125b3b40031"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.466471 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.493994 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.494258 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.494270 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.494282 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.494290 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16c800b6-4095-4676-ac1a-2125b3b40031-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595038 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-sg-core-conf-yaml\") pod \"d67f59d9-ce48-4306-b966-fdba768dc612\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595082 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-ceilometer-tls-certs\") pod \"d67f59d9-ce48-4306-b966-fdba768dc612\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595121 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-run-httpd\") pod \"d67f59d9-ce48-4306-b966-fdba768dc612\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595156 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wj7z\" (UniqueName: \"kubernetes.io/projected/d67f59d9-ce48-4306-b966-fdba768dc612-kube-api-access-5wj7z\") pod \"d67f59d9-ce48-4306-b966-fdba768dc612\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595221 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-scripts\") pod \"d67f59d9-ce48-4306-b966-fdba768dc612\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595313 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-log-httpd\") pod \"d67f59d9-ce48-4306-b966-fdba768dc612\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595389 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-combined-ca-bundle\") pod \"d67f59d9-ce48-4306-b966-fdba768dc612\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595485 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-config-data\") pod \"d67f59d9-ce48-4306-b966-fdba768dc612\" (UID: \"d67f59d9-ce48-4306-b966-fdba768dc612\") " Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595910 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d67f59d9-ce48-4306-b966-fdba768dc612" (UID: "d67f59d9-ce48-4306-b966-fdba768dc612"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.595979 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d67f59d9-ce48-4306-b966-fdba768dc612" (UID: "d67f59d9-ce48-4306-b966-fdba768dc612"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.596570 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.596594 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d67f59d9-ce48-4306-b966-fdba768dc612-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.599354 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-scripts" (OuterVolumeSpecName: "scripts") pod "d67f59d9-ce48-4306-b966-fdba768dc612" (UID: "d67f59d9-ce48-4306-b966-fdba768dc612"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.600767 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d67f59d9-ce48-4306-b966-fdba768dc612-kube-api-access-5wj7z" (OuterVolumeSpecName: "kube-api-access-5wj7z") pod "d67f59d9-ce48-4306-b966-fdba768dc612" (UID: "d67f59d9-ce48-4306-b966-fdba768dc612"). InnerVolumeSpecName "kube-api-access-5wj7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.629847 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d67f59d9-ce48-4306-b966-fdba768dc612" (UID: "d67f59d9-ce48-4306-b966-fdba768dc612"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.666704 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d67f59d9-ce48-4306-b966-fdba768dc612" (UID: "d67f59d9-ce48-4306-b966-fdba768dc612"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.697942 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wj7z\" (UniqueName: \"kubernetes.io/projected/d67f59d9-ce48-4306-b966-fdba768dc612-kube-api-access-5wj7z\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.697985 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.697999 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.698012 4733 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.711095 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d67f59d9-ce48-4306-b966-fdba768dc612" (UID: "d67f59d9-ce48-4306-b966-fdba768dc612"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.731310 4733 generic.go:334] "Generic (PLEG): container finished" podID="d67f59d9-ce48-4306-b966-fdba768dc612" containerID="1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75" exitCode=0 Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.731419 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.731468 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerDied","Data":"1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75"} Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.731520 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d67f59d9-ce48-4306-b966-fdba768dc612","Type":"ContainerDied","Data":"e4509d4bdeceeb929f056051733e62a9dae0bb7e5fe100aaf0d818bf3badeeb8"} Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.731538 4733 scope.go:117] "RemoveContainer" containerID="bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.732036 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-config-data" (OuterVolumeSpecName: "config-data") pod "d67f59d9-ce48-4306-b966-fdba768dc612" (UID: "d67f59d9-ce48-4306-b966-fdba768dc612"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.736934 4733 generic.go:334] "Generic (PLEG): container finished" podID="8906a11a-4c2c-4e38-865b-664be708e41e" containerID="4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288" exitCode=0 Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.736993 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8brz" event={"ID":"8906a11a-4c2c-4e38-865b-664be708e41e","Type":"ContainerDied","Data":"4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288"} Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.739821 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c86cb171-a0ab-4a6c-bf23-3b44e981976c","Type":"ContainerStarted","Data":"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d"} Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.745621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" event={"ID":"16c800b6-4095-4676-ac1a-2125b3b40031","Type":"ContainerDied","Data":"94f93239559b64b542a844ee5a5a4a77b9857fc47928b4c69e8be150f42b9ed0"} Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.745728 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d75688ddc-qbzmf" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.748126 4733 generic.go:334] "Generic (PLEG): container finished" podID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerID="3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac" exitCode=0 Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.748168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thnp8" event={"ID":"502fb5bc-ecaa-4a78-92fd-02b8eb138442","Type":"ContainerDied","Data":"3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac"} Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.762708 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.762107477 podStartE2EDuration="3.762107477s" podCreationTimestamp="2025-12-04 18:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:45.75768773 +0000 UTC m=+1427.713048776" watchObservedRunningTime="2025-12-04 18:02:45.762107477 +0000 UTC m=+1427.717468523" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.785313 4733 scope.go:117] "RemoveContainer" containerID="a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.799474 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.799505 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d67f59d9-ce48-4306-b966-fdba768dc612-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.837962 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-qbzmf"] Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.849419 4733 scope.go:117] "RemoveContainer" containerID="449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.857401 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d75688ddc-qbzmf"] Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.878529 4733 scope.go:117] "RemoveContainer" containerID="1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.905506 4733 scope.go:117] "RemoveContainer" containerID="bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d" Dec 04 18:02:45 crc kubenswrapper[4733]: E1204 18:02:45.906538 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d\": container with ID starting with bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d not found: ID does not exist" containerID="bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.906573 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d"} err="failed to get container status \"bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d\": rpc error: code = NotFound desc = could not find container \"bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d\": container with ID starting with bebbd9a69bbb012a23d8cd653f5461b0fee27bdfe80ff3d6fc84380ea8adfa0d not found: ID does not exist" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.906597 4733 scope.go:117] "RemoveContainer" containerID="a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2" Dec 04 18:02:45 crc kubenswrapper[4733]: E1204 18:02:45.927683 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2\": container with ID starting with a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2 not found: ID does not exist" containerID="a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.928245 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2"} err="failed to get container status \"a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2\": rpc error: code = NotFound desc = could not find container \"a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2\": container with ID starting with a1061a49e00bbfee73b7334e02610e942f57195a35dc8102ec1cb40ec46915e2 not found: ID does not exist" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.928410 4733 scope.go:117] "RemoveContainer" containerID="449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882" Dec 04 18:02:45 crc kubenswrapper[4733]: E1204 18:02:45.928990 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882\": container with ID starting with 449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882 not found: ID does not exist" containerID="449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.929030 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882"} err="failed to get container status \"449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882\": rpc error: code = NotFound desc = could not find container \"449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882\": container with ID starting with 449e0b88f52ada7222193ecd60d45b618b7e6bbafaf66eb475a13f26a9ce9882 not found: ID does not exist" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.929053 4733 scope.go:117] "RemoveContainer" containerID="1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75" Dec 04 18:02:45 crc kubenswrapper[4733]: E1204 18:02:45.931035 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75\": container with ID starting with 1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75 not found: ID does not exist" containerID="1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.931074 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75"} err="failed to get container status \"1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75\": rpc error: code = NotFound desc = could not find container \"1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75\": container with ID starting with 1ab7470b4578fc703716f027c84c8086cd9d2c9b7c4fac937d243b9c46db3f75 not found: ID does not exist" Dec 04 18:02:45 crc kubenswrapper[4733]: I1204 18:02:45.931101 4733 scope.go:117] "RemoveContainer" containerID="f32eabfab4f0ea2f0f6ea5778bd2054fd4fbd57466933574641025b23450044f" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.040865 4733 scope.go:117] "RemoveContainer" containerID="9063871b77f8aabc81ae5aa75dd5adc6cab49b2defefd7449f490514978e0928" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.105916 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.123404 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.169425 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:46 crc kubenswrapper[4733]: E1204 18:02:46.172020 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="proxy-httpd" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172053 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="proxy-httpd" Dec 04 18:02:46 crc kubenswrapper[4733]: E1204 18:02:46.172093 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="sg-core" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172101 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="sg-core" Dec 04 18:02:46 crc kubenswrapper[4733]: E1204 18:02:46.172114 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="ceilometer-notification-agent" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172124 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="ceilometer-notification-agent" Dec 04 18:02:46 crc kubenswrapper[4733]: E1204 18:02:46.172143 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c800b6-4095-4676-ac1a-2125b3b40031" containerName="dnsmasq-dns" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172150 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c800b6-4095-4676-ac1a-2125b3b40031" containerName="dnsmasq-dns" Dec 04 18:02:46 crc kubenswrapper[4733]: E1204 18:02:46.172166 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="ceilometer-central-agent" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172173 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="ceilometer-central-agent" Dec 04 18:02:46 crc kubenswrapper[4733]: E1204 18:02:46.172183 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c800b6-4095-4676-ac1a-2125b3b40031" containerName="init" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172191 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c800b6-4095-4676-ac1a-2125b3b40031" containerName="init" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172431 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c800b6-4095-4676-ac1a-2125b3b40031" containerName="dnsmasq-dns" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172451 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="proxy-httpd" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172466 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="sg-core" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172478 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="ceilometer-notification-agent" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.172488 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" containerName="ceilometer-central-agent" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.175024 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.178361 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.178662 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.179108 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.193390 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.317317 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-log-httpd\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.317368 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-config-data\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.317409 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.317437 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.317456 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkb4b\" (UniqueName: \"kubernetes.io/projected/90549281-a4f7-416e-92dd-cea0e94c9af7-kube-api-access-qkb4b\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.317489 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-scripts\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.317505 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.317543 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-run-httpd\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.348000 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16c800b6-4095-4676-ac1a-2125b3b40031" path="/var/lib/kubelet/pods/16c800b6-4095-4676-ac1a-2125b3b40031/volumes" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.348619 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d67f59d9-ce48-4306-b966-fdba768dc612" path="/var/lib/kubelet/pods/d67f59d9-ce48-4306-b966-fdba768dc612/volumes" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.419103 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-config-data\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.419185 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.419222 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.419246 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkb4b\" (UniqueName: \"kubernetes.io/projected/90549281-a4f7-416e-92dd-cea0e94c9af7-kube-api-access-qkb4b\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.419285 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-scripts\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.419315 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.419370 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-run-httpd\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.419576 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-log-httpd\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.420117 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-run-httpd\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.420176 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-log-httpd\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.425223 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-config-data\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.425857 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.426514 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-scripts\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.427595 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.428057 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.438088 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkb4b\" (UniqueName: \"kubernetes.io/projected/90549281-a4f7-416e-92dd-cea0e94c9af7-kube-api-access-qkb4b\") pod \"ceilometer-0\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.520124 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.763741 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8brz" event={"ID":"8906a11a-4c2c-4e38-865b-664be708e41e","Type":"ContainerStarted","Data":"eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361"} Dec 04 18:02:46 crc kubenswrapper[4733]: I1204 18:02:46.972560 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:02:47 crc kubenswrapper[4733]: I1204 18:02:47.777456 4733 generic.go:334] "Generic (PLEG): container finished" podID="8906a11a-4c2c-4e38-865b-664be708e41e" containerID="eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361" exitCode=0 Dec 04 18:02:47 crc kubenswrapper[4733]: I1204 18:02:47.777540 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8brz" event={"ID":"8906a11a-4c2c-4e38-865b-664be708e41e","Type":"ContainerDied","Data":"eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361"} Dec 04 18:02:47 crc kubenswrapper[4733]: I1204 18:02:47.781141 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="f1e31ad24991110558bf25e49f4521223e1cb18d3a3ee43198b8eff935ecc2a3" exitCode=0 Dec 04 18:02:47 crc kubenswrapper[4733]: I1204 18:02:47.781272 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"f1e31ad24991110558bf25e49f4521223e1cb18d3a3ee43198b8eff935ecc2a3"} Dec 04 18:02:47 crc kubenswrapper[4733]: I1204 18:02:47.781332 4733 scope.go:117] "RemoveContainer" containerID="f00595a749ca89ac07966111bfcf1cdeb8fcd3b039701228e808a5c55fac2e38" Dec 04 18:02:47 crc kubenswrapper[4733]: I1204 18:02:47.783312 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerStarted","Data":"595538b422dbb0c27c9c270478e9f80ea12d525054e7920843d616e22bc29c48"} Dec 04 18:02:48 crc kubenswrapper[4733]: I1204 18:02:48.823684 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thnp8" event={"ID":"502fb5bc-ecaa-4a78-92fd-02b8eb138442","Type":"ContainerStarted","Data":"3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984"} Dec 04 18:02:48 crc kubenswrapper[4733]: I1204 18:02:48.835611 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerStarted","Data":"43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864"} Dec 04 18:02:48 crc kubenswrapper[4733]: I1204 18:02:48.855721 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-thnp8" podStartSLOduration=3.249490031 podStartE2EDuration="8.85569704s" podCreationTimestamp="2025-12-04 18:02:40 +0000 UTC" firstStartedPulling="2025-12-04 18:02:42.578966856 +0000 UTC m=+1424.534327902" lastFinishedPulling="2025-12-04 18:02:48.185173855 +0000 UTC m=+1430.140534911" observedRunningTime="2025-12-04 18:02:48.845337395 +0000 UTC m=+1430.800698441" watchObservedRunningTime="2025-12-04 18:02:48.85569704 +0000 UTC m=+1430.811058086" Dec 04 18:02:49 crc kubenswrapper[4733]: I1204 18:02:49.845160 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7"} Dec 04 18:02:49 crc kubenswrapper[4733]: I1204 18:02:49.848151 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerStarted","Data":"aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a"} Dec 04 18:02:49 crc kubenswrapper[4733]: I1204 18:02:49.849649 4733 generic.go:334] "Generic (PLEG): container finished" podID="246f7a9e-8e85-4aa7-8bf4-44739e1d500d" containerID="9a63a5646f24f905ab643629b703d9fa8539f845df76e04dbee266c5247dbf9d" exitCode=0 Dec 04 18:02:49 crc kubenswrapper[4733]: I1204 18:02:49.849699 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ksgdb" event={"ID":"246f7a9e-8e85-4aa7-8bf4-44739e1d500d","Type":"ContainerDied","Data":"9a63a5646f24f905ab643629b703d9fa8539f845df76e04dbee266c5247dbf9d"} Dec 04 18:02:49 crc kubenswrapper[4733]: I1204 18:02:49.852522 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8brz" event={"ID":"8906a11a-4c2c-4e38-865b-664be708e41e","Type":"ContainerStarted","Data":"e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928"} Dec 04 18:02:49 crc kubenswrapper[4733]: I1204 18:02:49.888569 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f8brz" podStartSLOduration=3.390902342 podStartE2EDuration="6.888552084s" podCreationTimestamp="2025-12-04 18:02:43 +0000 UTC" firstStartedPulling="2025-12-04 18:02:44.801608643 +0000 UTC m=+1426.756969689" lastFinishedPulling="2025-12-04 18:02:48.299258375 +0000 UTC m=+1430.254619431" observedRunningTime="2025-12-04 18:02:49.883033508 +0000 UTC m=+1431.838394554" watchObservedRunningTime="2025-12-04 18:02:49.888552084 +0000 UTC m=+1431.843913130" Dec 04 18:02:50 crc kubenswrapper[4733]: I1204 18:02:50.863520 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerStarted","Data":"8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7"} Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.050558 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.051849 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.065379 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.195377 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.195431 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.265641 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.321887 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97chn\" (UniqueName: \"kubernetes.io/projected/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-kube-api-access-97chn\") pod \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.321964 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-combined-ca-bundle\") pod \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.322002 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-scripts\") pod \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.322114 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-config-data\") pod \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\" (UID: \"246f7a9e-8e85-4aa7-8bf4-44739e1d500d\") " Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.327340 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-kube-api-access-97chn" (OuterVolumeSpecName: "kube-api-access-97chn") pod "246f7a9e-8e85-4aa7-8bf4-44739e1d500d" (UID: "246f7a9e-8e85-4aa7-8bf4-44739e1d500d"). InnerVolumeSpecName "kube-api-access-97chn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.354426 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-scripts" (OuterVolumeSpecName: "scripts") pod "246f7a9e-8e85-4aa7-8bf4-44739e1d500d" (UID: "246f7a9e-8e85-4aa7-8bf4-44739e1d500d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.360612 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-config-data" (OuterVolumeSpecName: "config-data") pod "246f7a9e-8e85-4aa7-8bf4-44739e1d500d" (UID: "246f7a9e-8e85-4aa7-8bf4-44739e1d500d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.378927 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "246f7a9e-8e85-4aa7-8bf4-44739e1d500d" (UID: "246f7a9e-8e85-4aa7-8bf4-44739e1d500d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.425086 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.425126 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97chn\" (UniqueName: \"kubernetes.io/projected/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-kube-api-access-97chn\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.425138 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.425149 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/246f7a9e-8e85-4aa7-8bf4-44739e1d500d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.874647 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerStarted","Data":"cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4"} Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.875245 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.876187 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ksgdb" event={"ID":"246f7a9e-8e85-4aa7-8bf4-44739e1d500d","Type":"ContainerDied","Data":"4a0dc5ee94f7b67147c48e4958ce8050e758d62271c8fb681785fcc814368a24"} Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.876221 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ksgdb" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.876223 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a0dc5ee94f7b67147c48e4958ce8050e758d62271c8fb681785fcc814368a24" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.884917 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 18:02:51 crc kubenswrapper[4733]: I1204 18:02:51.903867 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.501392759 podStartE2EDuration="5.903845836s" podCreationTimestamp="2025-12-04 18:02:46 +0000 UTC" firstStartedPulling="2025-12-04 18:02:46.979650186 +0000 UTC m=+1428.935011252" lastFinishedPulling="2025-12-04 18:02:51.382103283 +0000 UTC m=+1433.337464329" observedRunningTime="2025-12-04 18:02:51.898016441 +0000 UTC m=+1433.853377517" watchObservedRunningTime="2025-12-04 18:02:51.903845836 +0000 UTC m=+1433.859206872" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.051627 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.052418 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerName="nova-api-log" containerID="cri-o://3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8" gracePeriod=30 Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.053049 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerName="nova-api-api" containerID="cri-o://fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d" gracePeriod=30 Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.073311 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.073584 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" containerName="nova-scheduler-scheduler" containerID="cri-o://756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d" gracePeriod=30 Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.115055 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.251186 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-thnp8" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="registry-server" probeResult="failure" output=< Dec 04 18:02:52 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 18:02:52 crc kubenswrapper[4733]: > Dec 04 18:02:52 crc kubenswrapper[4733]: E1204 18:02:52.468461 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc86cb171_a0ab_4a6c_bf23_3b44e981976c.slice/crio-conmon-fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d.scope\": RecentStats: unable to find data in memory cache]" Dec 04 18:02:52 crc kubenswrapper[4733]: E1204 18:02:52.513005 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:02:52 crc kubenswrapper[4733]: E1204 18:02:52.516582 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:02:52 crc kubenswrapper[4733]: E1204 18:02:52.518458 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:02:52 crc kubenswrapper[4733]: E1204 18:02:52.518519 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" containerName="nova-scheduler-scheduler" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.670106 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.859118 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-public-tls-certs\") pod \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.859187 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-config-data\") pod \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.859234 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96cwf\" (UniqueName: \"kubernetes.io/projected/c86cb171-a0ab-4a6c-bf23-3b44e981976c-kube-api-access-96cwf\") pod \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.859260 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-internal-tls-certs\") pod \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.859283 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86cb171-a0ab-4a6c-bf23-3b44e981976c-logs\") pod \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.859376 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-combined-ca-bundle\") pod \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\" (UID: \"c86cb171-a0ab-4a6c-bf23-3b44e981976c\") " Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.859898 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c86cb171-a0ab-4a6c-bf23-3b44e981976c-logs" (OuterVolumeSpecName: "logs") pod "c86cb171-a0ab-4a6c-bf23-3b44e981976c" (UID: "c86cb171-a0ab-4a6c-bf23-3b44e981976c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.877017 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c86cb171-a0ab-4a6c-bf23-3b44e981976c-kube-api-access-96cwf" (OuterVolumeSpecName: "kube-api-access-96cwf") pod "c86cb171-a0ab-4a6c-bf23-3b44e981976c" (UID: "c86cb171-a0ab-4a6c-bf23-3b44e981976c"). InnerVolumeSpecName "kube-api-access-96cwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.889416 4733 generic.go:334] "Generic (PLEG): container finished" podID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerID="fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d" exitCode=0 Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.889450 4733 generic.go:334] "Generic (PLEG): container finished" podID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerID="3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8" exitCode=143 Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.889500 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.889550 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c86cb171-a0ab-4a6c-bf23-3b44e981976c","Type":"ContainerDied","Data":"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d"} Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.889576 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c86cb171-a0ab-4a6c-bf23-3b44e981976c","Type":"ContainerDied","Data":"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8"} Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.889585 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c86cb171-a0ab-4a6c-bf23-3b44e981976c","Type":"ContainerDied","Data":"e240ce27cfe5cd8ea2ee788da03ef20acd32be270c77f3e99a57a0f1c2704168"} Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.889600 4733 scope.go:117] "RemoveContainer" containerID="fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.931030 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-config-data" (OuterVolumeSpecName: "config-data") pod "c86cb171-a0ab-4a6c-bf23-3b44e981976c" (UID: "c86cb171-a0ab-4a6c-bf23-3b44e981976c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.942045 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c86cb171-a0ab-4a6c-bf23-3b44e981976c" (UID: "c86cb171-a0ab-4a6c-bf23-3b44e981976c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.962138 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c86cb171-a0ab-4a6c-bf23-3b44e981976c" (UID: "c86cb171-a0ab-4a6c-bf23-3b44e981976c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.963814 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.963852 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.963865 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96cwf\" (UniqueName: \"kubernetes.io/projected/c86cb171-a0ab-4a6c-bf23-3b44e981976c-kube-api-access-96cwf\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.963877 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86cb171-a0ab-4a6c-bf23-3b44e981976c-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.963888 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:52 crc kubenswrapper[4733]: I1204 18:02:52.988197 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c86cb171-a0ab-4a6c-bf23-3b44e981976c" (UID: "c86cb171-a0ab-4a6c-bf23-3b44e981976c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.048937 4733 scope.go:117] "RemoveContainer" containerID="3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.068636 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86cb171-a0ab-4a6c-bf23-3b44e981976c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.073264 4733 scope.go:117] "RemoveContainer" containerID="fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d" Dec 04 18:02:53 crc kubenswrapper[4733]: E1204 18:02:53.073976 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d\": container with ID starting with fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d not found: ID does not exist" containerID="fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.074017 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d"} err="failed to get container status \"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d\": rpc error: code = NotFound desc = could not find container \"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d\": container with ID starting with fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d not found: ID does not exist" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.074045 4733 scope.go:117] "RemoveContainer" containerID="3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8" Dec 04 18:02:53 crc kubenswrapper[4733]: E1204 18:02:53.074300 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8\": container with ID starting with 3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8 not found: ID does not exist" containerID="3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.074327 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8"} err="failed to get container status \"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8\": rpc error: code = NotFound desc = could not find container \"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8\": container with ID starting with 3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8 not found: ID does not exist" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.074341 4733 scope.go:117] "RemoveContainer" containerID="fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.074708 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d"} err="failed to get container status \"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d\": rpc error: code = NotFound desc = could not find container \"fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d\": container with ID starting with fa86a3f583f7d674700f914c38ae2637d259ac8582b328db3b3df8a9c2dd595d not found: ID does not exist" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.074725 4733 scope.go:117] "RemoveContainer" containerID="3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.074975 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8"} err="failed to get container status \"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8\": rpc error: code = NotFound desc = could not find container \"3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8\": container with ID starting with 3a42fced528ab35732722de46a02977a7256641fdcc133576df62ac249b992b8 not found: ID does not exist" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.221852 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.233737 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.249271 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:53 crc kubenswrapper[4733]: E1204 18:02:53.249652 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerName="nova-api-log" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.249664 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerName="nova-api-log" Dec 04 18:02:53 crc kubenswrapper[4733]: E1204 18:02:53.249699 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerName="nova-api-api" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.249705 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerName="nova-api-api" Dec 04 18:02:53 crc kubenswrapper[4733]: E1204 18:02:53.249716 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246f7a9e-8e85-4aa7-8bf4-44739e1d500d" containerName="nova-manage" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.249722 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="246f7a9e-8e85-4aa7-8bf4-44739e1d500d" containerName="nova-manage" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.250473 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerName="nova-api-log" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.250501 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="246f7a9e-8e85-4aa7-8bf4-44739e1d500d" containerName="nova-manage" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.250511 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" containerName="nova-api-api" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.251437 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.253545 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.255049 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.255960 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.274503 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.374109 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f707420-7a54-4f2c-ac75-38e09d8dfb61-logs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.374185 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.374463 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.374509 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn5tm\" (UniqueName: \"kubernetes.io/projected/2f707420-7a54-4f2c-ac75-38e09d8dfb61-kube-api-access-xn5tm\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.374715 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-config-data\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.374835 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.476213 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f707420-7a54-4f2c-ac75-38e09d8dfb61-logs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.476286 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.476331 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.476364 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn5tm\" (UniqueName: \"kubernetes.io/projected/2f707420-7a54-4f2c-ac75-38e09d8dfb61-kube-api-access-xn5tm\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.476388 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-config-data\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.476554 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f707420-7a54-4f2c-ac75-38e09d8dfb61-logs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.476992 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.481619 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-config-data\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.485395 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.485767 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.494701 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn5tm\" (UniqueName: \"kubernetes.io/projected/2f707420-7a54-4f2c-ac75-38e09d8dfb61-kube-api-access-xn5tm\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.499140 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.573211 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.609309 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.610209 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.901030 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-log" containerID="cri-o://070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6" gracePeriod=30 Dec 04 18:02:53 crc kubenswrapper[4733]: I1204 18:02:53.901066 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-metadata" containerID="cri-o://7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b" gracePeriod=30 Dec 04 18:02:54 crc kubenswrapper[4733]: I1204 18:02:54.061044 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:02:54 crc kubenswrapper[4733]: W1204 18:02:54.061864 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f707420_7a54_4f2c_ac75_38e09d8dfb61.slice/crio-4b1d5f86b1d49b0f44bcb86cbce3bdb23b5b724a2ae48075989c0629e879f98e WatchSource:0}: Error finding container 4b1d5f86b1d49b0f44bcb86cbce3bdb23b5b724a2ae48075989c0629e879f98e: Status 404 returned error can't find the container with id 4b1d5f86b1d49b0f44bcb86cbce3bdb23b5b724a2ae48075989c0629e879f98e Dec 04 18:02:54 crc kubenswrapper[4733]: I1204 18:02:54.349366 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c86cb171-a0ab-4a6c-bf23-3b44e981976c" path="/var/lib/kubelet/pods/c86cb171-a0ab-4a6c-bf23-3b44e981976c/volumes" Dec 04 18:02:54 crc kubenswrapper[4733]: I1204 18:02:54.674012 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-f8brz" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="registry-server" probeResult="failure" output=< Dec 04 18:02:54 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 18:02:54 crc kubenswrapper[4733]: > Dec 04 18:02:54 crc kubenswrapper[4733]: I1204 18:02:54.914608 4733 generic.go:334] "Generic (PLEG): container finished" podID="c1acfe14-408d-4782-a419-9eb454c13de1" containerID="070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6" exitCode=143 Dec 04 18:02:54 crc kubenswrapper[4733]: I1204 18:02:54.914693 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1acfe14-408d-4782-a419-9eb454c13de1","Type":"ContainerDied","Data":"070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6"} Dec 04 18:02:54 crc kubenswrapper[4733]: I1204 18:02:54.917746 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f707420-7a54-4f2c-ac75-38e09d8dfb61","Type":"ContainerStarted","Data":"135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39"} Dec 04 18:02:54 crc kubenswrapper[4733]: I1204 18:02:54.917813 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f707420-7a54-4f2c-ac75-38e09d8dfb61","Type":"ContainerStarted","Data":"53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd"} Dec 04 18:02:54 crc kubenswrapper[4733]: I1204 18:02:54.917824 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f707420-7a54-4f2c-ac75-38e09d8dfb61","Type":"ContainerStarted","Data":"4b1d5f86b1d49b0f44bcb86cbce3bdb23b5b724a2ae48075989c0629e879f98e"} Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.368041 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.395969 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.395951343 podStartE2EDuration="3.395951343s" podCreationTimestamp="2025-12-04 18:02:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:54.94738319 +0000 UTC m=+1436.902744246" watchObservedRunningTime="2025-12-04 18:02:56.395951343 +0000 UTC m=+1438.351312389" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.442117 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-config-data\") pod \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.442261 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz8mc\" (UniqueName: \"kubernetes.io/projected/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-kube-api-access-jz8mc\") pod \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.442303 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-combined-ca-bundle\") pod \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\" (UID: \"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2\") " Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.447716 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-kube-api-access-jz8mc" (OuterVolumeSpecName: "kube-api-access-jz8mc") pod "cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" (UID: "cb3677c9-1c3f-4e2d-80d4-64fbe83fead2"). InnerVolumeSpecName "kube-api-access-jz8mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.468855 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-config-data" (OuterVolumeSpecName: "config-data") pod "cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" (UID: "cb3677c9-1c3f-4e2d-80d4-64fbe83fead2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.486071 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" (UID: "cb3677c9-1c3f-4e2d-80d4-64fbe83fead2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.544012 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.544048 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz8mc\" (UniqueName: \"kubernetes.io/projected/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-kube-api-access-jz8mc\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.544061 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.960286 4733 generic.go:334] "Generic (PLEG): container finished" podID="cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" containerID="756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d" exitCode=0 Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.960342 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2","Type":"ContainerDied","Data":"756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d"} Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.960368 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cb3677c9-1c3f-4e2d-80d4-64fbe83fead2","Type":"ContainerDied","Data":"9aee316e3e8f29a650a0abf47d49047f12930001f57b3a25aabf7d5ea434b90f"} Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.960387 4733 scope.go:117] "RemoveContainer" containerID="756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.960424 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.998467 4733 scope.go:117] "RemoveContainer" containerID="756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d" Dec 04 18:02:56 crc kubenswrapper[4733]: E1204 18:02:56.999303 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d\": container with ID starting with 756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d not found: ID does not exist" containerID="756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d" Dec 04 18:02:56 crc kubenswrapper[4733]: I1204 18:02:56.999345 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d"} err="failed to get container status \"756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d\": rpc error: code = NotFound desc = could not find container \"756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d\": container with ID starting with 756c0465a07a473b3644a573eea3fdb6f8b3ad66a78f37162f8b0d33ae82229d not found: ID does not exist" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.007846 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.021611 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.034558 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:57 crc kubenswrapper[4733]: E1204 18:02:57.035168 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" containerName="nova-scheduler-scheduler" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.035195 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" containerName="nova-scheduler-scheduler" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.035471 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" containerName="nova-scheduler-scheduler" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.036341 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.038597 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.046729 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.071027 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": read tcp 10.217.0.2:58924->10.217.0.197:8775: read: connection reset by peer" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.071286 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": read tcp 10.217.0.2:58926->10.217.0.197:8775: read: connection reset by peer" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.154342 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.154423 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6sq9\" (UniqueName: \"kubernetes.io/projected/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-kube-api-access-c6sq9\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.154781 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-config-data\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.264546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.264614 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6sq9\" (UniqueName: \"kubernetes.io/projected/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-kube-api-access-c6sq9\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.264712 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-config-data\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.277664 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.281425 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-config-data\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.296273 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6sq9\" (UniqueName: \"kubernetes.io/projected/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-kube-api-access-c6sq9\") pod \"nova-scheduler-0\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.367263 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.538000 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.571210 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1acfe14-408d-4782-a419-9eb454c13de1-logs\") pod \"c1acfe14-408d-4782-a419-9eb454c13de1\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.571285 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-nova-metadata-tls-certs\") pod \"c1acfe14-408d-4782-a419-9eb454c13de1\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.571368 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-combined-ca-bundle\") pod \"c1acfe14-408d-4782-a419-9eb454c13de1\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.571395 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-config-data\") pod \"c1acfe14-408d-4782-a419-9eb454c13de1\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.571431 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lm8w\" (UniqueName: \"kubernetes.io/projected/c1acfe14-408d-4782-a419-9eb454c13de1-kube-api-access-2lm8w\") pod \"c1acfe14-408d-4782-a419-9eb454c13de1\" (UID: \"c1acfe14-408d-4782-a419-9eb454c13de1\") " Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.576850 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1acfe14-408d-4782-a419-9eb454c13de1-logs" (OuterVolumeSpecName: "logs") pod "c1acfe14-408d-4782-a419-9eb454c13de1" (UID: "c1acfe14-408d-4782-a419-9eb454c13de1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.579214 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1acfe14-408d-4782-a419-9eb454c13de1-kube-api-access-2lm8w" (OuterVolumeSpecName: "kube-api-access-2lm8w") pod "c1acfe14-408d-4782-a419-9eb454c13de1" (UID: "c1acfe14-408d-4782-a419-9eb454c13de1"). InnerVolumeSpecName "kube-api-access-2lm8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.602485 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1acfe14-408d-4782-a419-9eb454c13de1" (UID: "c1acfe14-408d-4782-a419-9eb454c13de1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.642081 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-config-data" (OuterVolumeSpecName: "config-data") pod "c1acfe14-408d-4782-a419-9eb454c13de1" (UID: "c1acfe14-408d-4782-a419-9eb454c13de1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.654363 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c1acfe14-408d-4782-a419-9eb454c13de1" (UID: "c1acfe14-408d-4782-a419-9eb454c13de1"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.676615 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.676655 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.676665 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lm8w\" (UniqueName: \"kubernetes.io/projected/c1acfe14-408d-4782-a419-9eb454c13de1-kube-api-access-2lm8w\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.676677 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1acfe14-408d-4782-a419-9eb454c13de1-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.676685 4733 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1acfe14-408d-4782-a419-9eb454c13de1-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.883765 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:02:57 crc kubenswrapper[4733]: W1204 18:02:57.886394 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75c5521d_9c39_4fa7_88ad_11eb375c6d9a.slice/crio-4eab83c6136e711b66288a20c7055eea07d20bd55f58d36074cdb6a0d40733cc WatchSource:0}: Error finding container 4eab83c6136e711b66288a20c7055eea07d20bd55f58d36074cdb6a0d40733cc: Status 404 returned error can't find the container with id 4eab83c6136e711b66288a20c7055eea07d20bd55f58d36074cdb6a0d40733cc Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.976399 4733 generic.go:334] "Generic (PLEG): container finished" podID="c1acfe14-408d-4782-a419-9eb454c13de1" containerID="7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b" exitCode=0 Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.976441 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.976457 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1acfe14-408d-4782-a419-9eb454c13de1","Type":"ContainerDied","Data":"7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b"} Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.976837 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1acfe14-408d-4782-a419-9eb454c13de1","Type":"ContainerDied","Data":"72325e0e45e6cbf2d87a27dcb280f1e775075eec301e8b8c1a4da5cb6e190633"} Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.976861 4733 scope.go:117] "RemoveContainer" containerID="7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b" Dec 04 18:02:57 crc kubenswrapper[4733]: I1204 18:02:57.988482 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"75c5521d-9c39-4fa7-88ad-11eb375c6d9a","Type":"ContainerStarted","Data":"4eab83c6136e711b66288a20c7055eea07d20bd55f58d36074cdb6a0d40733cc"} Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.042551 4733 scope.go:117] "RemoveContainer" containerID="070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.049554 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.056752 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.070735 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:58 crc kubenswrapper[4733]: E1204 18:02:58.071128 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-log" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.071144 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-log" Dec 04 18:02:58 crc kubenswrapper[4733]: E1204 18:02:58.071157 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-metadata" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.071164 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-metadata" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.071379 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-metadata" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.071395 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" containerName="nova-metadata-log" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.072503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.081113 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.087346 4733 scope.go:117] "RemoveContainer" containerID="7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.087961 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.088097 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.089424 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.089539 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm4dx\" (UniqueName: \"kubernetes.io/projected/dd72019f-8a36-43af-8d73-7c3e804c1baa-kube-api-access-vm4dx\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.089566 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd72019f-8a36-43af-8d73-7c3e804c1baa-logs\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.089599 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-config-data\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.089628 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: E1204 18:02:58.092960 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b\": container with ID starting with 7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b not found: ID does not exist" containerID="7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.093023 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b"} err="failed to get container status \"7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b\": rpc error: code = NotFound desc = could not find container \"7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b\": container with ID starting with 7292efc1f0a501c58d885f4e6b45e608f9cf2f4b44eba42aaac8a0dad501726b not found: ID does not exist" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.093059 4733 scope.go:117] "RemoveContainer" containerID="070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6" Dec 04 18:02:58 crc kubenswrapper[4733]: E1204 18:02:58.096887 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6\": container with ID starting with 070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6 not found: ID does not exist" containerID="070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.096941 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6"} err="failed to get container status \"070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6\": rpc error: code = NotFound desc = could not find container \"070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6\": container with ID starting with 070e26608b04e403d801811026ed9e1308b0a5968ee37835742fe1688dab4db6 not found: ID does not exist" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.191299 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm4dx\" (UniqueName: \"kubernetes.io/projected/dd72019f-8a36-43af-8d73-7c3e804c1baa-kube-api-access-vm4dx\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.191349 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd72019f-8a36-43af-8d73-7c3e804c1baa-logs\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.191378 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-config-data\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.191419 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.191516 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.191808 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd72019f-8a36-43af-8d73-7c3e804c1baa-logs\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.196398 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-config-data\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.198439 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.202350 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.216038 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm4dx\" (UniqueName: \"kubernetes.io/projected/dd72019f-8a36-43af-8d73-7c3e804c1baa-kube-api-access-vm4dx\") pod \"nova-metadata-0\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.358484 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1acfe14-408d-4782-a419-9eb454c13de1" path="/var/lib/kubelet/pods/c1acfe14-408d-4782-a419-9eb454c13de1/volumes" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.362065 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb3677c9-1c3f-4e2d-80d4-64fbe83fead2" path="/var/lib/kubelet/pods/cb3677c9-1c3f-4e2d-80d4-64fbe83fead2/volumes" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.411580 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:02:58 crc kubenswrapper[4733]: I1204 18:02:58.909208 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:02:58 crc kubenswrapper[4733]: W1204 18:02:58.913637 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd72019f_8a36_43af_8d73_7c3e804c1baa.slice/crio-9a3d6a26497154ee9585e27e15e7ef56316cdacda43fca2b6de46e9d06f61e93 WatchSource:0}: Error finding container 9a3d6a26497154ee9585e27e15e7ef56316cdacda43fca2b6de46e9d06f61e93: Status 404 returned error can't find the container with id 9a3d6a26497154ee9585e27e15e7ef56316cdacda43fca2b6de46e9d06f61e93 Dec 04 18:02:59 crc kubenswrapper[4733]: I1204 18:02:59.007498 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd72019f-8a36-43af-8d73-7c3e804c1baa","Type":"ContainerStarted","Data":"9a3d6a26497154ee9585e27e15e7ef56316cdacda43fca2b6de46e9d06f61e93"} Dec 04 18:02:59 crc kubenswrapper[4733]: I1204 18:02:59.009400 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"75c5521d-9c39-4fa7-88ad-11eb375c6d9a","Type":"ContainerStarted","Data":"165c8851cb406979af0eb269716a0167a055f027eebbecd5ddc98f53e1a7d325"} Dec 04 18:02:59 crc kubenswrapper[4733]: I1204 18:02:59.043602 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.0435838139999998 podStartE2EDuration="2.043583814s" podCreationTimestamp="2025-12-04 18:02:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:02:59.021997771 +0000 UTC m=+1440.977358847" watchObservedRunningTime="2025-12-04 18:02:59.043583814 +0000 UTC m=+1440.998944860" Dec 04 18:03:00 crc kubenswrapper[4733]: I1204 18:03:00.031665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd72019f-8a36-43af-8d73-7c3e804c1baa","Type":"ContainerStarted","Data":"f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e"} Dec 04 18:03:00 crc kubenswrapper[4733]: I1204 18:03:00.032059 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd72019f-8a36-43af-8d73-7c3e804c1baa","Type":"ContainerStarted","Data":"39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb"} Dec 04 18:03:00 crc kubenswrapper[4733]: I1204 18:03:00.060937 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.060919958 podStartE2EDuration="2.060919958s" podCreationTimestamp="2025-12-04 18:02:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:03:00.057031914 +0000 UTC m=+1442.012392970" watchObservedRunningTime="2025-12-04 18:03:00.060919958 +0000 UTC m=+1442.016281024" Dec 04 18:03:01 crc kubenswrapper[4733]: I1204 18:03:01.246970 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:03:01 crc kubenswrapper[4733]: I1204 18:03:01.318580 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:03:01 crc kubenswrapper[4733]: I1204 18:03:01.490653 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-thnp8"] Dec 04 18:03:02 crc kubenswrapper[4733]: I1204 18:03:02.368385 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.057741 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-thnp8" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="registry-server" containerID="cri-o://3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984" gracePeriod=2 Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.411953 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.412330 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.574561 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.574604 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.603512 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.673200 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.700634 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvmk9\" (UniqueName: \"kubernetes.io/projected/502fb5bc-ecaa-4a78-92fd-02b8eb138442-kube-api-access-tvmk9\") pod \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.700864 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-utilities\") pod \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.700929 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-catalog-content\") pod \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\" (UID: \"502fb5bc-ecaa-4a78-92fd-02b8eb138442\") " Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.702870 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-utilities" (OuterVolumeSpecName: "utilities") pod "502fb5bc-ecaa-4a78-92fd-02b8eb138442" (UID: "502fb5bc-ecaa-4a78-92fd-02b8eb138442"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.709030 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/502fb5bc-ecaa-4a78-92fd-02b8eb138442-kube-api-access-tvmk9" (OuterVolumeSpecName: "kube-api-access-tvmk9") pod "502fb5bc-ecaa-4a78-92fd-02b8eb138442" (UID: "502fb5bc-ecaa-4a78-92fd-02b8eb138442"). InnerVolumeSpecName "kube-api-access-tvmk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.726345 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.804153 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.804234 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvmk9\" (UniqueName: \"kubernetes.io/projected/502fb5bc-ecaa-4a78-92fd-02b8eb138442-kube-api-access-tvmk9\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.814137 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "502fb5bc-ecaa-4a78-92fd-02b8eb138442" (UID: "502fb5bc-ecaa-4a78-92fd-02b8eb138442"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:03 crc kubenswrapper[4733]: I1204 18:03:03.906842 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/502fb5bc-ecaa-4a78-92fd-02b8eb138442-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.072077 4733 generic.go:334] "Generic (PLEG): container finished" podID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerID="3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984" exitCode=0 Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.072214 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-thnp8" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.072296 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thnp8" event={"ID":"502fb5bc-ecaa-4a78-92fd-02b8eb138442","Type":"ContainerDied","Data":"3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984"} Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.072362 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thnp8" event={"ID":"502fb5bc-ecaa-4a78-92fd-02b8eb138442","Type":"ContainerDied","Data":"a695e2d9962fb2460b3e992875a18ab2f2a5a22562cc2d17ea19b8d38d683b1c"} Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.072391 4733 scope.go:117] "RemoveContainer" containerID="3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.099555 4733 scope.go:117] "RemoveContainer" containerID="3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.122828 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-thnp8"] Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.132774 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-thnp8"] Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.153559 4733 scope.go:117] "RemoveContainer" containerID="d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.190188 4733 scope.go:117] "RemoveContainer" containerID="3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984" Dec 04 18:03:04 crc kubenswrapper[4733]: E1204 18:03:04.190782 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984\": container with ID starting with 3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984 not found: ID does not exist" containerID="3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.190840 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984"} err="failed to get container status \"3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984\": rpc error: code = NotFound desc = could not find container \"3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984\": container with ID starting with 3253144c4aafd7ce33d2952e9b1ce9904575ab2b2f96af216be199b776457984 not found: ID does not exist" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.190866 4733 scope.go:117] "RemoveContainer" containerID="3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac" Dec 04 18:03:04 crc kubenswrapper[4733]: E1204 18:03:04.192605 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac\": container with ID starting with 3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac not found: ID does not exist" containerID="3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.192635 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac"} err="failed to get container status \"3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac\": rpc error: code = NotFound desc = could not find container \"3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac\": container with ID starting with 3a50e2e6fed6fe4fc44b913093e61d42ccec15662c42f320cc105fa7409215ac not found: ID does not exist" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.192653 4733 scope.go:117] "RemoveContainer" containerID="d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5" Dec 04 18:03:04 crc kubenswrapper[4733]: E1204 18:03:04.192934 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5\": container with ID starting with d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5 not found: ID does not exist" containerID="d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.192966 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5"} err="failed to get container status \"d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5\": rpc error: code = NotFound desc = could not find container \"d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5\": container with ID starting with d4ed992ff384becaf1123a5f9a3622a9f14d2c4bd9081e041e7a9773b9ce67d5 not found: ID does not exist" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.347400 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" path="/var/lib/kubelet/pods/502fb5bc-ecaa-4a78-92fd-02b8eb138442/volumes" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.610049 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 18:03:04 crc kubenswrapper[4733]: I1204 18:03:04.610222 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 18:03:05 crc kubenswrapper[4733]: I1204 18:03:05.491655 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f8brz"] Dec 04 18:03:05 crc kubenswrapper[4733]: I1204 18:03:05.492064 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f8brz" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="registry-server" containerID="cri-o://e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928" gracePeriod=2 Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.021870 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.095857 4733 generic.go:334] "Generic (PLEG): container finished" podID="8906a11a-4c2c-4e38-865b-664be708e41e" containerID="e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928" exitCode=0 Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.095908 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8brz" event={"ID":"8906a11a-4c2c-4e38-865b-664be708e41e","Type":"ContainerDied","Data":"e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928"} Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.095930 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8brz" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.095941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8brz" event={"ID":"8906a11a-4c2c-4e38-865b-664be708e41e","Type":"ContainerDied","Data":"1c406b0ada12d9452c7de684f4e7740cd1d530de263ad1bfbc156c9b48f82b41"} Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.095960 4733 scope.go:117] "RemoveContainer" containerID="e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.118048 4733 scope.go:117] "RemoveContainer" containerID="eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.141629 4733 scope.go:117] "RemoveContainer" containerID="4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.145833 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-utilities\") pod \"8906a11a-4c2c-4e38-865b-664be708e41e\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.145919 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-catalog-content\") pod \"8906a11a-4c2c-4e38-865b-664be708e41e\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.146062 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgkjq\" (UniqueName: \"kubernetes.io/projected/8906a11a-4c2c-4e38-865b-664be708e41e-kube-api-access-mgkjq\") pod \"8906a11a-4c2c-4e38-865b-664be708e41e\" (UID: \"8906a11a-4c2c-4e38-865b-664be708e41e\") " Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.147003 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-utilities" (OuterVolumeSpecName: "utilities") pod "8906a11a-4c2c-4e38-865b-664be708e41e" (UID: "8906a11a-4c2c-4e38-865b-664be708e41e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.153496 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8906a11a-4c2c-4e38-865b-664be708e41e-kube-api-access-mgkjq" (OuterVolumeSpecName: "kube-api-access-mgkjq") pod "8906a11a-4c2c-4e38-865b-664be708e41e" (UID: "8906a11a-4c2c-4e38-865b-664be708e41e"). InnerVolumeSpecName "kube-api-access-mgkjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.196055 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8906a11a-4c2c-4e38-865b-664be708e41e" (UID: "8906a11a-4c2c-4e38-865b-664be708e41e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.227553 4733 scope.go:117] "RemoveContainer" containerID="e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928" Dec 04 18:03:06 crc kubenswrapper[4733]: E1204 18:03:06.227901 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928\": container with ID starting with e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928 not found: ID does not exist" containerID="e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.227934 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928"} err="failed to get container status \"e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928\": rpc error: code = NotFound desc = could not find container \"e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928\": container with ID starting with e2cf705b5d30bc37426df53cce99da8257b544b8f23e46fd097795d7dd125928 not found: ID does not exist" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.227962 4733 scope.go:117] "RemoveContainer" containerID="eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361" Dec 04 18:03:06 crc kubenswrapper[4733]: E1204 18:03:06.228231 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361\": container with ID starting with eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361 not found: ID does not exist" containerID="eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.228247 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361"} err="failed to get container status \"eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361\": rpc error: code = NotFound desc = could not find container \"eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361\": container with ID starting with eae2d05d8bd0ee1469935b7e97c9310e1873a2015ad74434e68be5e1bd559361 not found: ID does not exist" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.228259 4733 scope.go:117] "RemoveContainer" containerID="4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288" Dec 04 18:03:06 crc kubenswrapper[4733]: E1204 18:03:06.228641 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288\": container with ID starting with 4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288 not found: ID does not exist" containerID="4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.228666 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288"} err="failed to get container status \"4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288\": rpc error: code = NotFound desc = could not find container \"4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288\": container with ID starting with 4fcd583bdf70fba862fb884f74ebc49c5fce16c951c877f1dc8e7c798d5ee288 not found: ID does not exist" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.249287 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.249321 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgkjq\" (UniqueName: \"kubernetes.io/projected/8906a11a-4c2c-4e38-865b-664be708e41e-kube-api-access-mgkjq\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.249332 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8906a11a-4c2c-4e38-865b-664be708e41e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.417927 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f8brz"] Dec 04 18:03:06 crc kubenswrapper[4733]: I1204 18:03:06.425664 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f8brz"] Dec 04 18:03:07 crc kubenswrapper[4733]: I1204 18:03:07.368882 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 18:03:07 crc kubenswrapper[4733]: I1204 18:03:07.412289 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 18:03:08 crc kubenswrapper[4733]: I1204 18:03:08.143302 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 18:03:08 crc kubenswrapper[4733]: I1204 18:03:08.346016 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" path="/var/lib/kubelet/pods/8906a11a-4c2c-4e38-865b-664be708e41e/volumes" Dec 04 18:03:08 crc kubenswrapper[4733]: I1204 18:03:08.412728 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 18:03:08 crc kubenswrapper[4733]: I1204 18:03:08.412770 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 18:03:09 crc kubenswrapper[4733]: I1204 18:03:09.426998 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 18:03:09 crc kubenswrapper[4733]: I1204 18:03:09.427000 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 18:03:13 crc kubenswrapper[4733]: I1204 18:03:13.583688 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 18:03:13 crc kubenswrapper[4733]: I1204 18:03:13.584112 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 18:03:13 crc kubenswrapper[4733]: I1204 18:03:13.584564 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 18:03:13 crc kubenswrapper[4733]: I1204 18:03:13.584583 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 18:03:13 crc kubenswrapper[4733]: I1204 18:03:13.591460 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 18:03:13 crc kubenswrapper[4733]: I1204 18:03:13.601446 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 18:03:16 crc kubenswrapper[4733]: I1204 18:03:16.529478 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 18:03:18 crc kubenswrapper[4733]: I1204 18:03:18.419632 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 18:03:18 crc kubenswrapper[4733]: I1204 18:03:18.422585 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 18:03:18 crc kubenswrapper[4733]: I1204 18:03:18.431526 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 18:03:19 crc kubenswrapper[4733]: I1204 18:03:19.286400 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.729008 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.730904 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="7b19ede9-4e23-416d-bc7a-3674ed551c21" containerName="openstackclient" containerID="cri-o://ebae6d3dccdf5b42c07640f9dd249ddf0193a7662a64ce8f92ecd4ca8021c824" gracePeriod=2 Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.769592 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.836865 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.837112 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" containerName="ovn-northd" containerID="cri-o://11928a6e357a5e63f468c1daf6d2ee76e316437dd96f549ec172b0d953d18271" gracePeriod=30 Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.837494 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" containerName="openstack-network-exporter" containerID="cri-o://c3b464a1016d052759179bfb0e7253393babdea9ae2efb95d1c7ade0f58ff289" gracePeriod=30 Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.875142 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-rkmmv"] Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.896564 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-67rp6"] Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.896773 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-67rp6" podUID="bf38da41-b6c0-4551-9ce9-32146ebc2160" containerName="openstack-network-exporter" containerID="cri-o://afe99239cae679354f70187888627dd752b6e0f56cc59225d53f0c36a6b659f8" gracePeriod=30 Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.959509 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-j58xt"] Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.980401 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementee67-account-delete-sxlxx"] Dec 04 18:03:39 crc kubenswrapper[4733]: E1204 18:03:39.980826 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b19ede9-4e23-416d-bc7a-3674ed551c21" containerName="openstackclient" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.980839 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b19ede9-4e23-416d-bc7a-3674ed551c21" containerName="openstackclient" Dec 04 18:03:39 crc kubenswrapper[4733]: E1204 18:03:39.980857 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="extract-utilities" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.980863 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="extract-utilities" Dec 04 18:03:39 crc kubenswrapper[4733]: E1204 18:03:39.980881 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="registry-server" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.980887 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="registry-server" Dec 04 18:03:39 crc kubenswrapper[4733]: E1204 18:03:39.980898 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="registry-server" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.980905 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="registry-server" Dec 04 18:03:39 crc kubenswrapper[4733]: E1204 18:03:39.980914 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="extract-utilities" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.980921 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="extract-utilities" Dec 04 18:03:39 crc kubenswrapper[4733]: E1204 18:03:39.980935 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="extract-content" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.980943 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="extract-content" Dec 04 18:03:39 crc kubenswrapper[4733]: E1204 18:03:39.980952 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="extract-content" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.980959 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="extract-content" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.981118 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="502fb5bc-ecaa-4a78-92fd-02b8eb138442" containerName="registry-server" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.981137 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8906a11a-4c2c-4e38-865b-664be708e41e" containerName="registry-server" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.981154 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b19ede9-4e23-416d-bc7a-3674ed551c21" containerName="openstackclient" Dec 04 18:03:39 crc kubenswrapper[4733]: I1204 18:03:39.981815 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.039258 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78jw9\" (UniqueName: \"kubernetes.io/projected/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-kube-api-access-78jw9\") pod \"placementee67-account-delete-sxlxx\" (UID: \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\") " pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.039344 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts\") pod \"placementee67-account-delete-sxlxx\" (UID: \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\") " pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.041303 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.072077 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementee67-account-delete-sxlxx"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.104030 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance776f-account-delete-696d2"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.105292 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.140084 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance776f-account-delete-696d2"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.140622 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78jw9\" (UniqueName: \"kubernetes.io/projected/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-kube-api-access-78jw9\") pod \"placementee67-account-delete-sxlxx\" (UID: \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\") " pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.140719 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts\") pod \"placementee67-account-delete-sxlxx\" (UID: \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\") " pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.141439 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.141483 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data podName:5e605c62-64b4-4417-80bb-bc3387881f7a nodeName:}" failed. No retries permitted until 2025-12-04 18:03:40.641468168 +0000 UTC m=+1482.596829214 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data") pod "rabbitmq-cell1-server-0" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a") : configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.142162 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts\") pod \"placementee67-account-delete-sxlxx\" (UID: \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\") " pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.180901 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-2khcb"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.183584 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78jw9\" (UniqueName: \"kubernetes.io/projected/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-kube-api-access-78jw9\") pod \"placementee67-account-delete-sxlxx\" (UID: \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\") " pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.190074 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-2khcb"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.242810 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32514f48-c09d-4bfc-a407-bc0b454ca946-operator-scripts\") pod \"glance776f-account-delete-696d2\" (UID: \"32514f48-c09d-4bfc-a407-bc0b454ca946\") " pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.242905 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlgtv\" (UniqueName: \"kubernetes.io/projected/32514f48-c09d-4bfc-a407-bc0b454ca946-kube-api-access-jlgtv\") pod \"glance776f-account-delete-696d2\" (UID: \"32514f48-c09d-4bfc-a407-bc0b454ca946\") " pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.263363 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.276332 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder7119-account-delete-kzkjb"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.277554 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.295438 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder7119-account-delete-kzkjb"] Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.316343 4733 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-j58xt" message="Exiting ovn-controller (1) " Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.316378 4733 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-j58xt" podUID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" containerName="ovn-controller" containerID="cri-o://ab4787c533c2ddd15a894f803c59aaec9e2a588d19932dfde226bd69bf3dff66" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.316405 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-j58xt" podUID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" containerName="ovn-controller" containerID="cri-o://ab4787c533c2ddd15a894f803c59aaec9e2a588d19932dfde226bd69bf3dff66" gracePeriod=30 Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.354732 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts\") pod \"cinder7119-account-delete-kzkjb\" (UID: \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\") " pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.355104 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw9hg\" (UniqueName: \"kubernetes.io/projected/3f1c1320-5093-464a-a9a0-8a4cb668d70f-kube-api-access-pw9hg\") pod \"cinder7119-account-delete-kzkjb\" (UID: \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\") " pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.355244 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32514f48-c09d-4bfc-a407-bc0b454ca946-operator-scripts\") pod \"glance776f-account-delete-696d2\" (UID: \"32514f48-c09d-4bfc-a407-bc0b454ca946\") " pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.355343 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlgtv\" (UniqueName: \"kubernetes.io/projected/32514f48-c09d-4bfc-a407-bc0b454ca946-kube-api-access-jlgtv\") pod \"glance776f-account-delete-696d2\" (UID: \"32514f48-c09d-4bfc-a407-bc0b454ca946\") " pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.355867 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.355989 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data podName:1f55e362-d59c-4269-92c3-d5ca014a2ef1 nodeName:}" failed. No retries permitted until 2025-12-04 18:03:40.855974534 +0000 UTC m=+1482.811335660 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data") pod "rabbitmq-server-0" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1") : configmap "rabbitmq-config-data" not found Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.356562 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32514f48-c09d-4bfc-a407-bc0b454ca946-operator-scripts\") pod \"glance776f-account-delete-696d2\" (UID: \"32514f48-c09d-4bfc-a407-bc0b454ca946\") " pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.359392 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c53e91d-1256-408a-9b97-a048f425c441" path="/var/lib/kubelet/pods/8c53e91d-1256-408a-9b97-a048f425c441/volumes" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.396649 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican2aa9-account-delete-7bq8x"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.397877 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.418331 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlgtv\" (UniqueName: \"kubernetes.io/projected/32514f48-c09d-4bfc-a407-bc0b454ca946-kube-api-access-jlgtv\") pod \"glance776f-account-delete-696d2\" (UID: \"32514f48-c09d-4bfc-a407-bc0b454ca946\") " pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.421725 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican2aa9-account-delete-7bq8x"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.482697 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.487873 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stmgx\" (UniqueName: \"kubernetes.io/projected/31d98e38-ee17-4974-94e7-6495564940be-kube-api-access-stmgx\") pod \"barbican2aa9-account-delete-7bq8x\" (UID: \"31d98e38-ee17-4974-94e7-6495564940be\") " pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.488473 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d98e38-ee17-4974-94e7-6495564940be-operator-scripts\") pod \"barbican2aa9-account-delete-7bq8x\" (UID: \"31d98e38-ee17-4974-94e7-6495564940be\") " pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.502328 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts\") pod \"cinder7119-account-delete-kzkjb\" (UID: \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\") " pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.502434 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw9hg\" (UniqueName: \"kubernetes.io/projected/3f1c1320-5093-464a-a9a0-8a4cb668d70f-kube-api-access-pw9hg\") pod \"cinder7119-account-delete-kzkjb\" (UID: \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\") " pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.513413 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts\") pod \"cinder7119-account-delete-kzkjb\" (UID: \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\") " pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.513519 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-p2kxj"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.525713 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw9hg\" (UniqueName: \"kubernetes.io/projected/3f1c1320-5093-464a-a9a0-8a4cb668d70f-kube-api-access-pw9hg\") pod \"cinder7119-account-delete-kzkjb\" (UID: \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\") " pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.545284 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.582811 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-p2kxj"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.620609 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.625370 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron17fd-account-delete-pnhcm"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.629451 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stmgx\" (UniqueName: \"kubernetes.io/projected/31d98e38-ee17-4974-94e7-6495564940be-kube-api-access-stmgx\") pod \"barbican2aa9-account-delete-7bq8x\" (UID: \"31d98e38-ee17-4974-94e7-6495564940be\") " pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.636551 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d98e38-ee17-4974-94e7-6495564940be-operator-scripts\") pod \"barbican2aa9-account-delete-7bq8x\" (UID: \"31d98e38-ee17-4974-94e7-6495564940be\") " pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.637018 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d98e38-ee17-4974-94e7-6495564940be-operator-scripts\") pod \"barbican2aa9-account-delete-7bq8x\" (UID: \"31d98e38-ee17-4974-94e7-6495564940be\") " pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.639069 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron17fd-account-delete-pnhcm"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.639183 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.643719 4733 generic.go:334] "Generic (PLEG): container finished" podID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" containerID="ab4787c533c2ddd15a894f803c59aaec9e2a588d19932dfde226bd69bf3dff66" exitCode=0 Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.643834 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt" event={"ID":"3accb1a1-6ad9-42b9-9894-ba7b9e01d919","Type":"ContainerDied","Data":"ab4787c533c2ddd15a894f803c59aaec9e2a588d19932dfde226bd69bf3dff66"} Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.650786 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-67rp6_bf38da41-b6c0-4551-9ce9-32146ebc2160/openstack-network-exporter/0.log" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.650921 4733 generic.go:334] "Generic (PLEG): container finished" podID="bf38da41-b6c0-4551-9ce9-32146ebc2160" containerID="afe99239cae679354f70187888627dd752b6e0f56cc59225d53f0c36a6b659f8" exitCode=2 Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.650989 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-67rp6" event={"ID":"bf38da41-b6c0-4551-9ce9-32146ebc2160","Type":"ContainerDied","Data":"afe99239cae679354f70187888627dd752b6e0f56cc59225d53f0c36a6b659f8"} Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.668445 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stmgx\" (UniqueName: \"kubernetes.io/projected/31d98e38-ee17-4974-94e7-6495564940be-kube-api-access-stmgx\") pod \"barbican2aa9-account-delete-7bq8x\" (UID: \"31d98e38-ee17-4974-94e7-6495564940be\") " pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.671442 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-btknr"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.678277 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_229fa5f0-e006-4c04-8476-23a7842d292b/ovn-northd/0.log" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.678323 4733 generic.go:334] "Generic (PLEG): container finished" podID="229fa5f0-e006-4c04-8476-23a7842d292b" containerID="c3b464a1016d052759179bfb0e7253393babdea9ae2efb95d1c7ade0f58ff289" exitCode=2 Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.678340 4733 generic.go:334] "Generic (PLEG): container finished" podID="229fa5f0-e006-4c04-8476-23a7842d292b" containerID="11928a6e357a5e63f468c1daf6d2ee76e316437dd96f549ec172b0d953d18271" exitCode=143 Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.678362 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"229fa5f0-e006-4c04-8476-23a7842d292b","Type":"ContainerDied","Data":"c3b464a1016d052759179bfb0e7253393babdea9ae2efb95d1c7ade0f58ff289"} Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.678386 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"229fa5f0-e006-4c04-8476-23a7842d292b","Type":"ContainerDied","Data":"11928a6e357a5e63f468c1daf6d2ee76e316437dd96f549ec172b0d953d18271"} Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.688065 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-btknr"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.740661 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-operator-scripts\") pod \"neutron17fd-account-delete-pnhcm\" (UID: \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\") " pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.740829 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k8t2\" (UniqueName: \"kubernetes.io/projected/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-kube-api-access-5k8t2\") pod \"neutron17fd-account-delete-pnhcm\" (UID: \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\") " pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.741108 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.741178 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data podName:5e605c62-64b4-4417-80bb-bc3387881f7a nodeName:}" failed. No retries permitted until 2025-12-04 18:03:41.741160271 +0000 UTC m=+1483.696521317 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data") pod "rabbitmq-cell1-server-0" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a") : configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.824985 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.826322 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-gbxkg"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.846743 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-operator-scripts\") pod \"neutron17fd-account-delete-pnhcm\" (UID: \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\") " pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.847587 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-operator-scripts\") pod \"neutron17fd-account-delete-pnhcm\" (UID: \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\") " pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.847781 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k8t2\" (UniqueName: \"kubernetes.io/projected/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-kube-api-access-5k8t2\") pod \"neutron17fd-account-delete-pnhcm\" (UID: \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\") " pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.848291 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-gbxkg"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.874469 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d6df7889d-qkp4j"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.874692 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-d6df7889d-qkp4j" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerName="placement-log" containerID="cri-o://606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2" gracePeriod=30 Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.875130 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-d6df7889d-qkp4j" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerName="placement-api" containerID="cri-o://d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8" gracePeriod=30 Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.899990 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k8t2\" (UniqueName: \"kubernetes.io/projected/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-kube-api-access-5k8t2\") pod \"neutron17fd-account-delete-pnhcm\" (UID: \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\") " pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.933008 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.938310 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerName="openstack-network-exporter" containerID="cri-o://5c2bc4e3e18cd14bdc17ea6d5ff14205f792f1f10c8fef16751c74cb1ffb2a1e" gracePeriod=300 Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.949954 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 18:03:40 crc kubenswrapper[4733]: E1204 18:03:40.950015 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data podName:1f55e362-d59c-4269-92c3-d5ca014a2ef1 nodeName:}" failed. No retries permitted until 2025-12-04 18:03:41.949998457 +0000 UTC m=+1483.905359503 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data") pod "rabbitmq-server-0" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1") : configmap "rabbitmq-config-data" not found Dec 04 18:03:40 crc kubenswrapper[4733]: I1204 18:03:40.995252 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.084228 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi5c0f-account-delete-zk9rz"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.085464 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.087490 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi5c0f-account-delete-zk9rz"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.153788 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerName="ovsdbserver-nb" containerID="cri-o://995e1f7fa4114b63bde0864f52338990697384923eb1c3ee3773a27841eb999d" gracePeriod=300 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.172013 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.172326 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="openstack-network-exporter" containerID="cri-o://a2f20b96c0e76358892a2e4c8789ed0cb1a66ac87763e230752a2979798603be" gracePeriod=300 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.189992 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell06ac2-account-delete-rs267"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.191357 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.207089 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9t2z2"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.238685 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-9t2z2"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.256725 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6rss\" (UniqueName: \"kubernetes.io/projected/faf216c0-9cda-4f76-a4df-8855352db991-kube-api-access-z6rss\") pod \"novaapi5c0f-account-delete-zk9rz\" (UID: \"faf216c0-9cda-4f76-a4df-8855352db991\") " pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.256853 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/faf216c0-9cda-4f76-a4df-8855352db991-operator-scripts\") pod \"novaapi5c0f-account-delete-zk9rz\" (UID: \"faf216c0-9cda-4f76-a4df-8855352db991\") " pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.264512 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell06ac2-account-delete-rs267"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.283585 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2thjt"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.301764 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2thjt"] Dec 04 18:03:41 crc kubenswrapper[4733]: E1204 18:03:41.304979 4733 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 04 18:03:41 crc kubenswrapper[4733]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 04 18:03:41 crc kubenswrapper[4733]: + source /usr/local/bin/container-scripts/functions Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNBridge=br-int Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNRemote=tcp:localhost:6642 Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNEncapType=geneve Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNAvailabilityZones= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ EnableChassisAsGateway=true Dec 04 18:03:41 crc kubenswrapper[4733]: ++ PhysicalNetworks= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNHostName= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 04 18:03:41 crc kubenswrapper[4733]: ++ ovs_dir=/var/lib/openvswitch Dec 04 18:03:41 crc kubenswrapper[4733]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 04 18:03:41 crc kubenswrapper[4733]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 04 18:03:41 crc kubenswrapper[4733]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + sleep 0.5 Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + sleep 0.5 Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + cleanup_ovsdb_server_semaphore Dec 04 18:03:41 crc kubenswrapper[4733]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 18:03:41 crc kubenswrapper[4733]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 04 18:03:41 crc kubenswrapper[4733]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-rkmmv" message=< Dec 04 18:03:41 crc kubenswrapper[4733]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 04 18:03:41 crc kubenswrapper[4733]: + source /usr/local/bin/container-scripts/functions Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNBridge=br-int Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNRemote=tcp:localhost:6642 Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNEncapType=geneve Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNAvailabilityZones= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ EnableChassisAsGateway=true Dec 04 18:03:41 crc kubenswrapper[4733]: ++ PhysicalNetworks= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNHostName= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 04 18:03:41 crc kubenswrapper[4733]: ++ ovs_dir=/var/lib/openvswitch Dec 04 18:03:41 crc kubenswrapper[4733]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 04 18:03:41 crc kubenswrapper[4733]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 04 18:03:41 crc kubenswrapper[4733]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + sleep 0.5 Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + sleep 0.5 Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + cleanup_ovsdb_server_semaphore Dec 04 18:03:41 crc kubenswrapper[4733]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 18:03:41 crc kubenswrapper[4733]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 04 18:03:41 crc kubenswrapper[4733]: > Dec 04 18:03:41 crc kubenswrapper[4733]: E1204 18:03:41.305021 4733 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 04 18:03:41 crc kubenswrapper[4733]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 04 18:03:41 crc kubenswrapper[4733]: + source /usr/local/bin/container-scripts/functions Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNBridge=br-int Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNRemote=tcp:localhost:6642 Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNEncapType=geneve Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNAvailabilityZones= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ EnableChassisAsGateway=true Dec 04 18:03:41 crc kubenswrapper[4733]: ++ PhysicalNetworks= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ OVNHostName= Dec 04 18:03:41 crc kubenswrapper[4733]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 04 18:03:41 crc kubenswrapper[4733]: ++ ovs_dir=/var/lib/openvswitch Dec 04 18:03:41 crc kubenswrapper[4733]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 04 18:03:41 crc kubenswrapper[4733]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 04 18:03:41 crc kubenswrapper[4733]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + sleep 0.5 Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + sleep 0.5 Dec 04 18:03:41 crc kubenswrapper[4733]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 18:03:41 crc kubenswrapper[4733]: + cleanup_ovsdb_server_semaphore Dec 04 18:03:41 crc kubenswrapper[4733]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 18:03:41 crc kubenswrapper[4733]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 04 18:03:41 crc kubenswrapper[4733]: > pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" containerID="cri-o://102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.305055 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" containerID="cri-o://102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" gracePeriod=29 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.330256 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="ovsdbserver-sb" containerID="cri-o://f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052" gracePeriod=300 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.347655 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348299 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-server" containerID="cri-o://afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348348 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-server" containerID="cri-o://e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348368 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-auditor" containerID="cri-o://cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348448 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-updater" containerID="cri-o://e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348486 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-updater" containerID="cri-o://aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348493 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-auditor" containerID="cri-o://33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348504 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-replicator" containerID="cri-o://1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348533 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-expirer" containerID="cri-o://b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348531 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-auditor" containerID="cri-o://a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348552 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="swift-recon-cron" containerID="cri-o://5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348559 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-replicator" containerID="cri-o://da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348491 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-reaper" containerID="cri-o://13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348588 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="rsync" containerID="cri-o://92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348239 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-server" containerID="cri-o://1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.348540 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-replicator" containerID="cri-o://e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.358830 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6rss\" (UniqueName: \"kubernetes.io/projected/faf216c0-9cda-4f76-a4df-8855352db991-kube-api-access-z6rss\") pod \"novaapi5c0f-account-delete-zk9rz\" (UID: \"faf216c0-9cda-4f76-a4df-8855352db991\") " pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.358912 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts\") pod \"novacell06ac2-account-delete-rs267\" (UID: \"32b94636-cf95-42f9-8589-e38b550053af\") " pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.358949 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/faf216c0-9cda-4f76-a4df-8855352db991-operator-scripts\") pod \"novaapi5c0f-account-delete-zk9rz\" (UID: \"faf216c0-9cda-4f76-a4df-8855352db991\") " pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.358978 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hrxp\" (UniqueName: \"kubernetes.io/projected/32b94636-cf95-42f9-8589-e38b550053af-kube-api-access-4hrxp\") pod \"novacell06ac2-account-delete-rs267\" (UID: \"32b94636-cf95-42f9-8589-e38b550053af\") " pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.363225 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/faf216c0-9cda-4f76-a4df-8855352db991-operator-scripts\") pod \"novaapi5c0f-account-delete-zk9rz\" (UID: \"faf216c0-9cda-4f76-a4df-8855352db991\") " pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.363282 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-zs7pg"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.363497 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" podUID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" containerName="dnsmasq-dns" containerID="cri-o://7ba1c201c7cef13fb8704e43d271cf5b0bf7479dc2ba8022c5b90de90fe6784e" gracePeriod=10 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.373675 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.374500 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerName="glance-log" containerID="cri-o://5fbdddb85e7e76ccd5911f7a283bd09d9eae7deb69d967d864ce48da4e7ac1be" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.374984 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerName="glance-httpd" containerID="cri-o://6560f5a95ae3f9b652dd82bb257bd795655e85bb32c02059e6da7c32d0b36e95" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.421312 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6rss\" (UniqueName: \"kubernetes.io/projected/faf216c0-9cda-4f76-a4df-8855352db991-kube-api-access-z6rss\") pod \"novaapi5c0f-account-delete-zk9rz\" (UID: \"faf216c0-9cda-4f76-a4df-8855352db991\") " pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.430539 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" containerID="cri-o://8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" gracePeriod=29 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.468495 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts\") pod \"novacell06ac2-account-delete-rs267\" (UID: \"32b94636-cf95-42f9-8589-e38b550053af\") " pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.468560 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hrxp\" (UniqueName: \"kubernetes.io/projected/32b94636-cf95-42f9-8589-e38b550053af-kube-api-access-4hrxp\") pod \"novacell06ac2-account-delete-rs267\" (UID: \"32b94636-cf95-42f9-8589-e38b550053af\") " pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.479592 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts\") pod \"novacell06ac2-account-delete-rs267\" (UID: \"32b94636-cf95-42f9-8589-e38b550053af\") " pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.495898 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.496287 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-log" containerID="cri-o://63006d63457f2e071c6d109c86778556fbf386e34e8d0116bd25dc63bd13509f" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.496749 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-httpd" containerID="cri-o://cbcf529109c0c4af1bdf871c346e94fb3c2778251cd7b3b1d1b5864a31229cb3" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.508727 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.509209 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api-log" containerID="cri-o://d366e224d098b7e5af2be8bd08ebc779d71f9239fa811ef64c92bd18e26bace0" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.509641 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api" containerID="cri-o://5cd076e1396593ece03d523ffdfe9f53cc72e35677e8aab21b5111c2bcce8fa1" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.531254 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hrxp\" (UniqueName: \"kubernetes.io/projected/32b94636-cf95-42f9-8589-e38b550053af-kube-api-access-4hrxp\") pod \"novacell06ac2-account-delete-rs267\" (UID: \"32b94636-cf95-42f9-8589-e38b550053af\") " pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:03:41 crc kubenswrapper[4733]: E1204 18:03:41.548286 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052 is running failed: container process not found" containerID="f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 18:03:41 crc kubenswrapper[4733]: E1204 18:03:41.548699 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052 is running failed: container process not found" containerID="f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 18:03:41 crc kubenswrapper[4733]: E1204 18:03:41.549108 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052 is running failed: container process not found" containerID="f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 18:03:41 crc kubenswrapper[4733]: E1204 18:03:41.549134 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="ovsdbserver-sb" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.583286 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.592020 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerName="cinder-scheduler" containerID="cri-o://9aa9ac8aec81ac8172b07681bf3447f72d402bc17fafe55ed65c2f79900cd3fb" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.593039 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerName="probe" containerID="cri-o://3044b3de56fc15c0557352a235e182ec8eaffbf427385fd0dad2161e82e28cd2" gracePeriod=30 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.638501 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.728912 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerName="rabbitmq" containerID="cri-o://d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d" gracePeriod=604800 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.756453 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.791516 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.799939 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xv4z2"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.850323 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xv4z2"] Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.853561 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j46r9\" (UniqueName: \"kubernetes.io/projected/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-kube-api-access-j46r9\") pod \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.854012 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-ovn-controller-tls-certs\") pod \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.856117 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-combined-ca-bundle\") pod \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.856166 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-scripts\") pod \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.856230 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-log-ovn\") pod \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.856245 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run-ovn\") pod \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.856262 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run\") pod \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\" (UID: \"3accb1a1-6ad9-42b9-9894-ba7b9e01d919\") " Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.857503 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "3accb1a1-6ad9-42b9-9894-ba7b9e01d919" (UID: "3accb1a1-6ad9-42b9-9894-ba7b9e01d919"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.867761 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run" (OuterVolumeSpecName: "var-run") pod "3accb1a1-6ad9-42b9-9894-ba7b9e01d919" (UID: "3accb1a1-6ad9-42b9-9894-ba7b9e01d919"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.868978 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "3accb1a1-6ad9-42b9-9894-ba7b9e01d919" (UID: "3accb1a1-6ad9-42b9-9894-ba7b9e01d919"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.871538 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-scripts" (OuterVolumeSpecName: "scripts") pod "3accb1a1-6ad9-42b9-9894-ba7b9e01d919" (UID: "3accb1a1-6ad9-42b9-9894-ba7b9e01d919"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.883766 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-ksgdb"] Dec 04 18:03:41 crc kubenswrapper[4733]: E1204 18:03:41.895716 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:41 crc kubenswrapper[4733]: E1204 18:03:41.896072 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data podName:5e605c62-64b4-4417-80bb-bc3387881f7a nodeName:}" failed. No retries permitted until 2025-12-04 18:03:43.896034686 +0000 UTC m=+1485.851395722 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data") pod "rabbitmq-cell1-server-0" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a") : configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.902074 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-kube-api-access-j46r9" (OuterVolumeSpecName: "kube-api-access-j46r9") pod "3accb1a1-6ad9-42b9-9894-ba7b9e01d919" (UID: "3accb1a1-6ad9-42b9-9894-ba7b9e01d919"). InnerVolumeSpecName "kube-api-access-j46r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.907166 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919079 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j46r9\" (UniqueName: \"kubernetes.io/projected/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-kube-api-access-j46r9\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919109 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919129 4733 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919142 4733 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919152 4733 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919319 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ae61093e-d566-48cb-acc2-dd88c4512b8e/ovsdbserver-sb/0.log" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919388 4733 generic.go:334] "Generic (PLEG): container finished" podID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerID="a2f20b96c0e76358892a2e4c8789ed0cb1a66ac87763e230752a2979798603be" exitCode=2 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919406 4733 generic.go:334] "Generic (PLEG): container finished" podID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerID="f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052" exitCode=143 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919508 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ae61093e-d566-48cb-acc2-dd88c4512b8e","Type":"ContainerDied","Data":"a2f20b96c0e76358892a2e4c8789ed0cb1a66ac87763e230752a2979798603be"} Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.919557 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ae61093e-d566-48cb-acc2-dd88c4512b8e","Type":"ContainerDied","Data":"f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052"} Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.923786 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-67rp6_bf38da41-b6c0-4551-9ce9-32146ebc2160/openstack-network-exporter/0.log" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.924026 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-67rp6" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.927547 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_229fa5f0-e006-4c04-8476-23a7842d292b/ovn-northd/0.log" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.927710 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.935454 4733 generic.go:334] "Generic (PLEG): container finished" podID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerID="5fbdddb85e7e76ccd5911f7a283bd09d9eae7deb69d967d864ce48da4e7ac1be" exitCode=143 Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.936414 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d","Type":"ContainerDied","Data":"5fbdddb85e7e76ccd5911f7a283bd09d9eae7deb69d967d864ce48da4e7ac1be"} Dec 04 18:03:41 crc kubenswrapper[4733]: I1204 18:03:41.998910 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-ksgdb"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.020727 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84697bbbd9-9fltw"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.021512 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-84697bbbd9-9fltw" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerName="neutron-api" containerID="cri-o://5a96bd3fd257dc89a4415f7d6d660060c72c788e7620add6273636ed06ed77b1" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.021968 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-84697bbbd9-9fltw" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerName="neutron-httpd" containerID="cri-o://04763b048c7378462084dba4301a110daf2dd03bc1d5722e8ceee83b2170fcc9" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.043230 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-rundir\") pod \"229fa5f0-e006-4c04-8476-23a7842d292b\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.043286 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-combined-ca-bundle\") pod \"229fa5f0-e006-4c04-8476-23a7842d292b\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.043321 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-metrics-certs-tls-certs\") pod \"bf38da41-b6c0-4551-9ce9-32146ebc2160\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.043363 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-combined-ca-bundle\") pod \"bf38da41-b6c0-4551-9ce9-32146ebc2160\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.043389 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-scripts\") pod \"229fa5f0-e006-4c04-8476-23a7842d292b\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.043655 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl4mv\" (UniqueName: \"kubernetes.io/projected/bf38da41-b6c0-4551-9ce9-32146ebc2160-kube-api-access-tl4mv\") pod \"bf38da41-b6c0-4551-9ce9-32146ebc2160\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.043696 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf38da41-b6c0-4551-9ce9-32146ebc2160-config\") pod \"bf38da41-b6c0-4551-9ce9-32146ebc2160\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.043983 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-northd-tls-certs\") pod \"229fa5f0-e006-4c04-8476-23a7842d292b\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.044058 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovs-rundir\") pod \"bf38da41-b6c0-4551-9ce9-32146ebc2160\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.044110 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.044136 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovn-rundir\") pod \"bf38da41-b6c0-4551-9ce9-32146ebc2160\" (UID: \"bf38da41-b6c0-4551-9ce9-32146ebc2160\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.044222 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7sn8\" (UniqueName: \"kubernetes.io/projected/229fa5f0-e006-4c04-8476-23a7842d292b-kube-api-access-r7sn8\") pod \"229fa5f0-e006-4c04-8476-23a7842d292b\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.044242 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-config\") pod \"229fa5f0-e006-4c04-8476-23a7842d292b\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.044272 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-metrics-certs-tls-certs\") pod \"229fa5f0-e006-4c04-8476-23a7842d292b\" (UID: \"229fa5f0-e006-4c04-8476-23a7842d292b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: E1204 18:03:42.045415 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 18:03:42 crc kubenswrapper[4733]: E1204 18:03:42.045478 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data podName:1f55e362-d59c-4269-92c3-d5ca014a2ef1 nodeName:}" failed. No retries permitted until 2025-12-04 18:03:44.045457704 +0000 UTC m=+1486.000818760 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data") pod "rabbitmq-server-0" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1") : configmap "rabbitmq-config-data" not found Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.046016 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "229fa5f0-e006-4c04-8476-23a7842d292b" (UID: "229fa5f0-e006-4c04-8476-23a7842d292b"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.052628 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "bf38da41-b6c0-4551-9ce9-32146ebc2160" (UID: "bf38da41-b6c0-4551-9ce9-32146ebc2160"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.053297 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-config" (OuterVolumeSpecName: "config") pod "229fa5f0-e006-4c04-8476-23a7842d292b" (UID: "229fa5f0-e006-4c04-8476-23a7842d292b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.053342 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "bf38da41-b6c0-4551-9ce9-32146ebc2160" (UID: "bf38da41-b6c0-4551-9ce9-32146ebc2160"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.092303 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf38da41-b6c0-4551-9ce9-32146ebc2160-config" (OuterVolumeSpecName: "config") pod "bf38da41-b6c0-4551-9ce9-32146ebc2160" (UID: "bf38da41-b6c0-4551-9ce9-32146ebc2160"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.106670 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3accb1a1-6ad9-42b9-9894-ba7b9e01d919" (UID: "3accb1a1-6ad9-42b9-9894-ba7b9e01d919"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.110461 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/229fa5f0-e006-4c04-8476-23a7842d292b-kube-api-access-r7sn8" (OuterVolumeSpecName: "kube-api-access-r7sn8") pod "229fa5f0-e006-4c04-8476-23a7842d292b" (UID: "229fa5f0-e006-4c04-8476-23a7842d292b"). InnerVolumeSpecName "kube-api-access-r7sn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.110689 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-scripts" (OuterVolumeSpecName: "scripts") pod "229fa5f0-e006-4c04-8476-23a7842d292b" (UID: "229fa5f0-e006-4c04-8476-23a7842d292b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.110770 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf38da41-b6c0-4551-9ce9-32146ebc2160-kube-api-access-tl4mv" (OuterVolumeSpecName: "kube-api-access-tl4mv") pod "bf38da41-b6c0-4551-9ce9-32146ebc2160" (UID: "bf38da41-b6c0-4551-9ce9-32146ebc2160"). InnerVolumeSpecName "kube-api-access-tl4mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112779 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112818 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112825 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112834 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112841 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112848 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112854 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112860 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112868 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112873 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112950 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.112988 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.113006 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.113015 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.113023 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.113033 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.113042 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.113050 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.113058 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.125518 4733 generic.go:334] "Generic (PLEG): container finished" podID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.125587 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rkmmv" event={"ID":"0cf011d5-1c33-4ec6-8b1f-d1216a46d066","Type":"ContainerDied","Data":"102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.126095 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "229fa5f0-e006-4c04-8476-23a7842d292b" (UID: "229fa5f0-e006-4c04-8476-23a7842d292b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.128663 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bae2e2ab-b6af-4d4d-8288-9f23050553a0/ovsdbserver-nb/0.log" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.128712 4733 generic.go:334] "Generic (PLEG): container finished" podID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerID="5c2bc4e3e18cd14bdc17ea6d5ff14205f792f1f10c8fef16751c74cb1ffb2a1e" exitCode=2 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.128730 4733 generic.go:334] "Generic (PLEG): container finished" podID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerID="995e1f7fa4114b63bde0864f52338990697384923eb1c3ee3773a27841eb999d" exitCode=143 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.128771 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bae2e2ab-b6af-4d4d-8288-9f23050553a0","Type":"ContainerDied","Data":"5c2bc4e3e18cd14bdc17ea6d5ff14205f792f1f10c8fef16751c74cb1ffb2a1e"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.128823 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bae2e2ab-b6af-4d4d-8288-9f23050553a0","Type":"ContainerDied","Data":"995e1f7fa4114b63bde0864f52338990697384923eb1c3ee3773a27841eb999d"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.133376 4733 generic.go:334] "Generic (PLEG): container finished" podID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerID="606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2" exitCode=143 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.133427 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6df7889d-qkp4j" event={"ID":"f575fafd-64f0-4401-82f1-f4a6a33b132e","Type":"ContainerDied","Data":"606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.136657 4733 generic.go:334] "Generic (PLEG): container finished" podID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" containerID="7ba1c201c7cef13fb8704e43d271cf5b0bf7479dc2ba8022c5b90de90fe6784e" exitCode=0 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.136727 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" event={"ID":"7446c8ed-a97b-41a1-ae84-8f1eed8b203b","Type":"ContainerDied","Data":"7ba1c201c7cef13fb8704e43d271cf5b0bf7479dc2ba8022c5b90de90fe6784e"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.143172 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-j58xt" event={"ID":"3accb1a1-6ad9-42b9-9894-ba7b9e01d919","Type":"ContainerDied","Data":"46ddd1e4cb00852ea370b1849f0d03314b82549b44b2ce83ff01416b7b9abc9a"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.143221 4733 scope.go:117] "RemoveContainer" containerID="ab4787c533c2ddd15a894f803c59aaec9e2a588d19932dfde226bd69bf3dff66" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.143366 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147194 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147223 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147231 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147244 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147253 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl4mv\" (UniqueName: \"kubernetes.io/projected/bf38da41-b6c0-4551-9ce9-32146ebc2160-kube-api-access-tl4mv\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147261 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf38da41-b6c0-4551-9ce9-32146ebc2160-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147269 4733 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147276 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/bf38da41-b6c0-4551-9ce9-32146ebc2160-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147289 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7sn8\" (UniqueName: \"kubernetes.io/projected/229fa5f0-e006-4c04-8476-23a7842d292b-kube-api-access-r7sn8\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.147298 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229fa5f0-e006-4c04-8476-23a7842d292b-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.161971 4733 generic.go:334] "Generic (PLEG): container finished" podID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerID="d366e224d098b7e5af2be8bd08ebc779d71f9239fa811ef64c92bd18e26bace0" exitCode=143 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.162020 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12b58db1-9574-4081-a3c1-2853f003fa8f","Type":"ContainerDied","Data":"d366e224d098b7e5af2be8bd08ebc779d71f9239fa811ef64c92bd18e26bace0"} Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.165673 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-676cf9f69-6mwxl"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.166036 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-676cf9f69-6mwxl" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerName="barbican-worker-log" containerID="cri-o://a6c4a59c75c37a61feea92a8e8114a64759adb31de0de0bbb5f0d5725e5b60bd" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.166201 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-676cf9f69-6mwxl" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerName="barbican-worker" containerID="cri-o://594133e5106de2f14d436d478b5b6fa76670e3bcd1bb710de01b002b98c8b79a" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.182917 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-75746f8666-mw9nb"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.183168 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerName="barbican-keystone-listener-log" containerID="cri-o://b2b953dbdaed812d6f01b224d3af9906a0d0da35f8c6eaeed44b254649426640" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.183589 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerName="barbican-keystone-listener" containerID="cri-o://52c25be5cf28de218adebb584f7e6fe433e56d68d1b191f31c36c67cb6d7c1a1" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.192990 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7f757c77dd-bzp5x"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.193628 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7f757c77dd-bzp5x" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api-log" containerID="cri-o://1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.194818 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7f757c77dd-bzp5x" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api" containerID="cri-o://a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.199069 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf38da41-b6c0-4551-9ce9-32146ebc2160" (UID: "bf38da41-b6c0-4551-9ce9-32146ebc2160"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.208347 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerName="rabbitmq" containerID="cri-o://98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25" gracePeriod=604800 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.226142 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-59945f4865-9zsr4"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.226373 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-59945f4865-9zsr4" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerName="proxy-httpd" containerID="cri-o://7a464410f79ac8c4ad9cc078a988fa37fc20c807e7742c3550f9f6de916082bb" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.226777 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-59945f4865-9zsr4" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerName="proxy-server" containerID="cri-o://f130a979912194eb23dc33f38388f755ef2b29dc69498aa6ea5c028aff72030b" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.251537 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.259314 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.273973 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.274468 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-log" containerID="cri-o://53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.275331 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-api" containerID="cri-o://135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.299092 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "3accb1a1-6ad9-42b9-9894-ba7b9e01d919" (UID: "3accb1a1-6ad9-42b9-9894-ba7b9e01d919"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.318040 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.318278 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-log" containerID="cri-o://39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.318656 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-metadata" containerID="cri-o://f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.346083 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ae61093e-d566-48cb-acc2-dd88c4512b8e/ovsdbserver-sb/0.log" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.346154 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.346493 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-ztq89"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.351292 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "229fa5f0-e006-4c04-8476-23a7842d292b" (UID: "229fa5f0-e006-4c04-8476-23a7842d292b"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.365051 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.365086 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3accb1a1-6ad9-42b9-9894-ba7b9e01d919-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.391017 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "bf38da41-b6c0-4551-9ce9-32146ebc2160" (UID: "bf38da41-b6c0-4551-9ce9-32146ebc2160"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.406970 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="246f7a9e-8e85-4aa7-8bf4-44739e1d500d" path="/var/lib/kubelet/pods/246f7a9e-8e85-4aa7-8bf4-44739e1d500d/volumes" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.407636 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3590b958-1ae5-46f6-9664-d8efdec482b2" path="/var/lib/kubelet/pods/3590b958-1ae5-46f6-9664-d8efdec482b2/volumes" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.408178 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bcc7f96-8d09-4d23-8e53-02b679d7851a" path="/var/lib/kubelet/pods/3bcc7f96-8d09-4d23-8e53-02b679d7851a/volumes" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.412648 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3be601db-3698-4272-8005-083d5b93cbfd" path="/var/lib/kubelet/pods/3be601db-3698-4272-8005-083d5b93cbfd/volumes" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.413495 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="553ecb2b-6910-4f64-ac13-93fa1afe5137" path="/var/lib/kubelet/pods/553ecb2b-6910-4f64-ac13-93fa1afe5137/volumes" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.415903 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77bd0d5e-6bf1-4257-b30e-795688595c1c" path="/var/lib/kubelet/pods/77bd0d5e-6bf1-4257-b30e-795688595c1c/volumes" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.416415 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="857cacd9-91d0-4a3a-92b8-73669fd75dfc" path="/var/lib/kubelet/pods/857cacd9-91d0-4a3a-92b8-73669fd75dfc/volumes" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.435156 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "229fa5f0-e006-4c04-8476-23a7842d292b" (UID: "229fa5f0-e006-4c04-8476-23a7842d292b"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.436045 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-ztq89"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.436089 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-a59c-account-create-update-wkkp2"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.436103 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-a59c-account-create-update-wkkp2"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.436114 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t5z7h"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.436123 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t5z7h"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.436133 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.436146 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8fxb9"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.436317 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="5eecf779-672e-4469-a134-9ecb0ef1bdde" containerName="nova-cell0-conductor-conductor" containerID="cri-o://96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.452752 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8fxb9"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.453249 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.453420 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="839d7537-ff12-4ee6-b36c-b192f130a6e4" containerName="nova-cell1-conductor-conductor" containerID="cri-o://bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.465811 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-metrics-certs-tls-certs\") pod \"ae61093e-d566-48cb-acc2-dd88c4512b8e\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.465913 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpqgh\" (UniqueName: \"kubernetes.io/projected/ae61093e-d566-48cb-acc2-dd88c4512b8e-kube-api-access-fpqgh\") pod \"ae61093e-d566-48cb-acc2-dd88c4512b8e\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466085 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdb-rundir\") pod \"ae61093e-d566-48cb-acc2-dd88c4512b8e\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466109 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-combined-ca-bundle\") pod \"ae61093e-d566-48cb-acc2-dd88c4512b8e\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466131 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-config\") pod \"ae61093e-d566-48cb-acc2-dd88c4512b8e\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466156 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ae61093e-d566-48cb-acc2-dd88c4512b8e\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466195 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdbserver-sb-tls-certs\") pod \"ae61093e-d566-48cb-acc2-dd88c4512b8e\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466222 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-scripts\") pod \"ae61093e-d566-48cb-acc2-dd88c4512b8e\" (UID: \"ae61093e-d566-48cb-acc2-dd88c4512b8e\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466613 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/229fa5f0-e006-4c04-8476-23a7842d292b-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466625 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf38da41-b6c0-4551-9ce9-32146ebc2160-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.466979 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.467160 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4047083c-6418-4814-8b24-9407494d0303" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.468144 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "ae61093e-d566-48cb-acc2-dd88c4512b8e" (UID: "ae61093e-d566-48cb-acc2-dd88c4512b8e"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.468584 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-config" (OuterVolumeSpecName: "config") pod "ae61093e-d566-48cb-acc2-dd88c4512b8e" (UID: "ae61093e-d566-48cb-acc2-dd88c4512b8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.469003 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-scripts" (OuterVolumeSpecName: "scripts") pod "ae61093e-d566-48cb-acc2-dd88c4512b8e" (UID: "ae61093e-d566-48cb-acc2-dd88c4512b8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.479316 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "ae61093e-d566-48cb-acc2-dd88c4512b8e" (UID: "ae61093e-d566-48cb-acc2-dd88c4512b8e"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.481654 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.481862 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="75c5521d-9c39-4fa7-88ad-11eb375c6d9a" containerName="nova-scheduler-scheduler" containerID="cri-o://165c8851cb406979af0eb269716a0167a055f027eebbecd5ddc98f53e1a7d325" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.490305 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae61093e-d566-48cb-acc2-dd88c4512b8e-kube-api-access-fpqgh" (OuterVolumeSpecName: "kube-api-access-fpqgh") pod "ae61093e-d566-48cb-acc2-dd88c4512b8e" (UID: "ae61093e-d566-48cb-acc2-dd88c4512b8e"). InnerVolumeSpecName "kube-api-access-fpqgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.527911 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae61093e-d566-48cb-acc2-dd88c4512b8e" (UID: "ae61093e-d566-48cb-acc2-dd88c4512b8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.539759 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" containerName="galera" containerID="cri-o://51718f0a98bc4f12648ae0c8e161095d9435571807a28e34ed23fb1ce1e33217" gracePeriod=30 Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.554950 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "ae61093e-d566-48cb-acc2-dd88c4512b8e" (UID: "ae61093e-d566-48cb-acc2-dd88c4512b8e"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.568842 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.568864 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpqgh\" (UniqueName: \"kubernetes.io/projected/ae61093e-d566-48cb-acc2-dd88c4512b8e-kube-api-access-fpqgh\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.568873 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.568882 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.568891 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae61093e-d566-48cb-acc2-dd88c4512b8e-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.568910 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.568918 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.579935 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "ae61093e-d566-48cb-acc2-dd88c4512b8e" (UID: "ae61093e-d566-48cb-acc2-dd88c4512b8e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.602777 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.645077 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bae2e2ab-b6af-4d4d-8288-9f23050553a0/ovsdbserver-nb/0.log" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.645351 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.667445 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.670276 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.670299 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae61093e-d566-48cb-acc2-dd88c4512b8e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.771888 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdb-rundir\") pod \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.771933 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-nb\") pod \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.771953 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.771979 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-combined-ca-bundle\") pod \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772042 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-scripts\") pod \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772060 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-metrics-certs-tls-certs\") pod \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772082 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96wk2\" (UniqueName: \"kubernetes.io/projected/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-kube-api-access-96wk2\") pod \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772099 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-sb\") pod \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772120 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghkf6\" (UniqueName: \"kubernetes.io/projected/bae2e2ab-b6af-4d4d-8288-9f23050553a0-kube-api-access-ghkf6\") pod \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772149 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-config\") pod \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772179 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdbserver-nb-tls-certs\") pod \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\" (UID: \"bae2e2ab-b6af-4d4d-8288-9f23050553a0\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772193 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-svc\") pod \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772240 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-swift-storage-0\") pod \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.772276 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-config\") pod \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\" (UID: \"7446c8ed-a97b-41a1-ae84-8f1eed8b203b\") " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.775285 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "bae2e2ab-b6af-4d4d-8288-9f23050553a0" (UID: "bae2e2ab-b6af-4d4d-8288-9f23050553a0"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.778699 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-config" (OuterVolumeSpecName: "config") pod "bae2e2ab-b6af-4d4d-8288-9f23050553a0" (UID: "bae2e2ab-b6af-4d4d-8288-9f23050553a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.779173 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-scripts" (OuterVolumeSpecName: "scripts") pod "bae2e2ab-b6af-4d4d-8288-9f23050553a0" (UID: "bae2e2ab-b6af-4d4d-8288-9f23050553a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.799987 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "bae2e2ab-b6af-4d4d-8288-9f23050553a0" (UID: "bae2e2ab-b6af-4d4d-8288-9f23050553a0"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.799997 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae2e2ab-b6af-4d4d-8288-9f23050553a0-kube-api-access-ghkf6" (OuterVolumeSpecName: "kube-api-access-ghkf6") pod "bae2e2ab-b6af-4d4d-8288-9f23050553a0" (UID: "bae2e2ab-b6af-4d4d-8288-9f23050553a0"). InnerVolumeSpecName "kube-api-access-ghkf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.803026 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-kube-api-access-96wk2" (OuterVolumeSpecName: "kube-api-access-96wk2") pod "7446c8ed-a97b-41a1-ae84-8f1eed8b203b" (UID: "7446c8ed-a97b-41a1-ae84-8f1eed8b203b"). InnerVolumeSpecName "kube-api-access-96wk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.873996 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.874037 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.874050 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.874059 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96wk2\" (UniqueName: \"kubernetes.io/projected/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-kube-api-access-96wk2\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.874069 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghkf6\" (UniqueName: \"kubernetes.io/projected/bae2e2ab-b6af-4d4d-8288-9f23050553a0-kube-api-access-ghkf6\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.874077 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae2e2ab-b6af-4d4d-8288-9f23050553a0-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.951210 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-config" (OuterVolumeSpecName: "config") pod "7446c8ed-a97b-41a1-ae84-8f1eed8b203b" (UID: "7446c8ed-a97b-41a1-ae84-8f1eed8b203b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:42 crc kubenswrapper[4733]: I1204 18:03:42.976058 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.007001 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.007934 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7446c8ed-a97b-41a1-ae84-8f1eed8b203b" (UID: "7446c8ed-a97b-41a1-ae84-8f1eed8b203b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.009747 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7446c8ed-a97b-41a1-ae84-8f1eed8b203b" (UID: "7446c8ed-a97b-41a1-ae84-8f1eed8b203b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.062926 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7446c8ed-a97b-41a1-ae84-8f1eed8b203b" (UID: "7446c8ed-a97b-41a1-ae84-8f1eed8b203b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.067989 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7446c8ed-a97b-41a1-ae84-8f1eed8b203b" (UID: "7446c8ed-a97b-41a1-ae84-8f1eed8b203b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.071848 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bae2e2ab-b6af-4d4d-8288-9f23050553a0" (UID: "bae2e2ab-b6af-4d4d-8288-9f23050553a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.077728 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementee67-account-delete-sxlxx"] Dec 04 18:03:43 crc kubenswrapper[4733]: E1204 18:03:43.081066 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:43 crc kubenswrapper[4733]: E1204 18:03:43.085348 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.085494 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican2aa9-account-delete-7bq8x"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.093404 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder7119-account-delete-kzkjb"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.094077 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.094161 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.094305 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.094362 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.094412 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.094460 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7446c8ed-a97b-41a1-ae84-8f1eed8b203b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: E1204 18:03:43.096557 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:43 crc kubenswrapper[4733]: E1204 18:03:43.096622 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="5eecf779-672e-4469-a134-9ecb0ef1bdde" containerName="nova-cell0-conductor-conductor" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.125474 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance776f-account-delete-696d2"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.147675 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron17fd-account-delete-pnhcm"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.148967 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "bae2e2ab-b6af-4d4d-8288-9f23050553a0" (UID: "bae2e2ab-b6af-4d4d-8288-9f23050553a0"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: W1204 18:03:43.159583 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08cdf00b_cf60_4e3a_9283_833e3a8d95bf.slice/crio-358da79aa4080332b38fbb6621db2dfd0b2f29bed19e6891cfa753d0cf58ca24 WatchSource:0}: Error finding container 358da79aa4080332b38fbb6621db2dfd0b2f29bed19e6891cfa753d0cf58ca24: Status 404 returned error can't find the container with id 358da79aa4080332b38fbb6621db2dfd0b2f29bed19e6891cfa753d0cf58ca24 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.192200 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" event={"ID":"7446c8ed-a97b-41a1-ae84-8f1eed8b203b","Type":"ContainerDied","Data":"e7892397ee71ab39db17b4f9ce076b0a9c65abc9b24ca5e76b41e706b70aa675"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.192249 4733 scope.go:117] "RemoveContainer" containerID="7ba1c201c7cef13fb8704e43d271cf5b0bf7479dc2ba8022c5b90de90fe6784e" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.192309 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf4ff87b5-zs7pg" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.200960 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.207415 4733 generic.go:334] "Generic (PLEG): container finished" podID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerID="a6c4a59c75c37a61feea92a8e8114a64759adb31de0de0bbb5f0d5725e5b60bd" exitCode=143 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.207558 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-676cf9f69-6mwxl" event={"ID":"87f72b89-cea6-4243-8ec6-e1264b4901c3","Type":"ContainerDied","Data":"a6c4a59c75c37a61feea92a8e8114a64759adb31de0de0bbb5f0d5725e5b60bd"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.213604 4733 generic.go:334] "Generic (PLEG): container finished" podID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerID="39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb" exitCode=143 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.213674 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd72019f-8a36-43af-8d73-7c3e804c1baa","Type":"ContainerDied","Data":"39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.230145 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bae2e2ab-b6af-4d4d-8288-9f23050553a0/ovsdbserver-nb/0.log" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.230513 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bae2e2ab-b6af-4d4d-8288-9f23050553a0","Type":"ContainerDied","Data":"f70927bc9d53b516e4d204c03b0fb379b1c49f82a18c06f3a57df744e6f37932"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.230624 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.242215 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "bae2e2ab-b6af-4d4d-8288-9f23050553a0" (UID: "bae2e2ab-b6af-4d4d-8288-9f23050553a0"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.243196 4733 generic.go:334] "Generic (PLEG): container finished" podID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerID="53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd" exitCode=143 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.243259 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f707420-7a54-4f2c-ac75-38e09d8dfb61","Type":"ContainerDied","Data":"53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.276743 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.276770 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.276777 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.276785 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.276875 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.276911 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.276926 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.276936 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.282475 4733 generic.go:334] "Generic (PLEG): container finished" podID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerID="b2b953dbdaed812d6f01b224d3af9906a0d0da35f8c6eaeed44b254649426640" exitCode=143 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.282605 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" event={"ID":"c1de9dcb-fad2-485d-b59b-14646f9ddfdd","Type":"ContainerDied","Data":"b2b953dbdaed812d6f01b224d3af9906a0d0da35f8c6eaeed44b254649426640"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.299492 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ae61093e-d566-48cb-acc2-dd88c4512b8e/ovsdbserver-sb/0.log" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.299573 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ae61093e-d566-48cb-acc2-dd88c4512b8e","Type":"ContainerDied","Data":"162cddb3f0e641814c81d2b1c26c22e815c3678b6df38c946878e1635dfee126"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.299672 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.303280 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae2e2ab-b6af-4d4d-8288-9f23050553a0-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.305989 4733 generic.go:334] "Generic (PLEG): container finished" podID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerID="f130a979912194eb23dc33f38388f755ef2b29dc69498aa6ea5c028aff72030b" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.306012 4733 generic.go:334] "Generic (PLEG): container finished" podID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerID="7a464410f79ac8c4ad9cc078a988fa37fc20c807e7742c3550f9f6de916082bb" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.306052 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59945f4865-9zsr4" event={"ID":"d78c8335-c2e4-4c49-8b77-98fec3cab751","Type":"ContainerDied","Data":"f130a979912194eb23dc33f38388f755ef2b29dc69498aa6ea5c028aff72030b"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.306076 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59945f4865-9zsr4" event={"ID":"d78c8335-c2e4-4c49-8b77-98fec3cab751","Type":"ContainerDied","Data":"7a464410f79ac8c4ad9cc078a988fa37fc20c807e7742c3550f9f6de916082bb"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.313148 4733 generic.go:334] "Generic (PLEG): container finished" podID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerID="1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d" exitCode=143 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.313209 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f757c77dd-bzp5x" event={"ID":"3c5f9b26-6c76-4af7-a811-d7d763f74ed6","Type":"ContainerDied","Data":"1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.316172 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-67rp6_bf38da41-b6c0-4551-9ce9-32146ebc2160/openstack-network-exporter/0.log" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.316536 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-67rp6" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.317664 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-67rp6" event={"ID":"bf38da41-b6c0-4551-9ce9-32146ebc2160","Type":"ContainerDied","Data":"be3a9bbf8278f801466aea9f70acb74e4cc5979dd71988bb769e38914b206d82"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.320649 4733 generic.go:334] "Generic (PLEG): container finished" podID="7b19ede9-4e23-416d-bc7a-3674ed551c21" containerID="ebae6d3dccdf5b42c07640f9dd249ddf0193a7662a64ce8f92ecd4ca8021c824" exitCode=137 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.320702 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd2552a52e606992fd0d93f83f3c9eacf2686b424ed0dbe674142bb023634cdc" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.324177 4733 generic.go:334] "Generic (PLEG): container finished" podID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerID="63006d63457f2e071c6d109c86778556fbf386e34e8d0116bd25dc63bd13509f" exitCode=143 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.324225 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c8b0056-e347-4f24-aeea-20680c33bf06","Type":"ContainerDied","Data":"63006d63457f2e071c6d109c86778556fbf386e34e8d0116bd25dc63bd13509f"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.326886 4733 generic.go:334] "Generic (PLEG): container finished" podID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerID="3044b3de56fc15c0557352a235e182ec8eaffbf427385fd0dad2161e82e28cd2" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.326905 4733 generic.go:334] "Generic (PLEG): container finished" podID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerID="9aa9ac8aec81ac8172b07681bf3447f72d402bc17fafe55ed65c2f79900cd3fb" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.326940 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f7df4d1e-71e7-4c40-a3e4-27b484799d09","Type":"ContainerDied","Data":"3044b3de56fc15c0557352a235e182ec8eaffbf427385fd0dad2161e82e28cd2"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.326958 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f7df4d1e-71e7-4c40-a3e4-27b484799d09","Type":"ContainerDied","Data":"9aa9ac8aec81ac8172b07681bf3447f72d402bc17fafe55ed65c2f79900cd3fb"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.326968 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f7df4d1e-71e7-4c40-a3e4-27b484799d09","Type":"ContainerDied","Data":"496c76265d05efa68b400a73f274d401dcb2d8c058678bef6ad325d66387d6a8"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.326977 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="496c76265d05efa68b400a73f274d401dcb2d8c058678bef6ad325d66387d6a8" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.328653 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron17fd-account-delete-pnhcm" event={"ID":"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc","Type":"ContainerStarted","Data":"909db43f49a047e2b4c76332e149c57629f691f23f6b6d720cf2e6be9ed4d49a"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.330107 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder7119-account-delete-kzkjb" event={"ID":"3f1c1320-5093-464a-a9a0-8a4cb668d70f","Type":"ContainerStarted","Data":"dd7e5dd3b4dd495bfcc89bbbb5553e1aa540defc806d3f337d11380f89e63395"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.333621 4733 generic.go:334] "Generic (PLEG): container finished" podID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerID="04763b048c7378462084dba4301a110daf2dd03bc1d5722e8ceee83b2170fcc9" exitCode=0 Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.333695 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84697bbbd9-9fltw" event={"ID":"4a9d9f43-f063-47fc-83b9-213c6ab06ef4","Type":"ContainerDied","Data":"04763b048c7378462084dba4301a110daf2dd03bc1d5722e8ceee83b2170fcc9"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.334603 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican2aa9-account-delete-7bq8x" event={"ID":"31d98e38-ee17-4974-94e7-6495564940be","Type":"ContainerStarted","Data":"6d53b5e0b7d789cf409282116070d1e8faedf4c50aabf9aaab81037888ca54c4"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.337397 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_229fa5f0-e006-4c04-8476-23a7842d292b/ovn-northd/0.log" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.337447 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"229fa5f0-e006-4c04-8476-23a7842d292b","Type":"ContainerDied","Data":"aa67deef74c1ebb3aa5e2511dbccc8d8adfecb54e13529613f7271be932a343d"} Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.337532 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.392530 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.393508 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi5c0f-account-delete-zk9rz"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.411948 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell06ac2-account-delete-rs267"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.444645 4733 scope.go:117] "RemoveContainer" containerID="40bce759ec7e8d43134a5240b984af64caf430aa9f304d8430db8fd918772895" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.507525 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-combined-ca-bundle\") pod \"7b19ede9-4e23-416d-bc7a-3674ed551c21\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.507669 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config-secret\") pod \"7b19ede9-4e23-416d-bc7a-3674ed551c21\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.507731 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config\") pod \"7b19ede9-4e23-416d-bc7a-3674ed551c21\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.507833 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk2q8\" (UniqueName: \"kubernetes.io/projected/7b19ede9-4e23-416d-bc7a-3674ed551c21-kube-api-access-zk2q8\") pod \"7b19ede9-4e23-416d-bc7a-3674ed551c21\" (UID: \"7b19ede9-4e23-416d-bc7a-3674ed551c21\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.514673 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b19ede9-4e23-416d-bc7a-3674ed551c21-kube-api-access-zk2q8" (OuterVolumeSpecName: "kube-api-access-zk2q8") pod "7b19ede9-4e23-416d-bc7a-3674ed551c21" (UID: "7b19ede9-4e23-416d-bc7a-3674ed551c21"). InnerVolumeSpecName "kube-api-access-zk2q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.543072 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.559411 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7b19ede9-4e23-416d-bc7a-3674ed551c21" (UID: "7b19ede9-4e23-416d-bc7a-3674ed551c21"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.583009 4733 scope.go:117] "RemoveContainer" containerID="5c2bc4e3e18cd14bdc17ea6d5ff14205f792f1f10c8fef16751c74cb1ffb2a1e" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.595086 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b19ede9-4e23-416d-bc7a-3674ed551c21" (UID: "7b19ede9-4e23-416d-bc7a-3674ed551c21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.610939 4733 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.610974 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk2q8\" (UniqueName: \"kubernetes.io/projected/7b19ede9-4e23-416d-bc7a-3674ed551c21-kube-api-access-zk2q8\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.610984 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.626239 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.647035 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.659456 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-zs7pg"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.668698 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf4ff87b5-zs7pg"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.673108 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7b19ede9-4e23-416d-bc7a-3674ed551c21" (UID: "7b19ede9-4e23-416d-bc7a-3674ed551c21"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.675176 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-67rp6"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.684257 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-67rp6"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.702414 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.705326 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.707434 4733 scope.go:117] "RemoveContainer" containerID="995e1f7fa4114b63bde0864f52338990697384923eb1c3ee3773a27841eb999d" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.711653 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7df4d1e-71e7-4c40-a3e4-27b484799d09-etc-machine-id\") pod \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.711731 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data-custom\") pod \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.711841 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj8fc\" (UniqueName: \"kubernetes.io/projected/f7df4d1e-71e7-4c40-a3e4-27b484799d09-kube-api-access-qj8fc\") pod \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.711863 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-combined-ca-bundle\") pod \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.711949 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-scripts\") pod \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.711981 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data\") pod \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\" (UID: \"f7df4d1e-71e7-4c40-a3e4-27b484799d09\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.712341 4733 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7b19ede9-4e23-416d-bc7a-3674ed551c21-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.715467 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.715880 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7df4d1e-71e7-4c40-a3e4-27b484799d09-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f7df4d1e-71e7-4c40-a3e4-27b484799d09" (UID: "f7df4d1e-71e7-4c40-a3e4-27b484799d09"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.716993 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7df4d1e-71e7-4c40-a3e4-27b484799d09-kube-api-access-qj8fc" (OuterVolumeSpecName: "kube-api-access-qj8fc") pod "f7df4d1e-71e7-4c40-a3e4-27b484799d09" (UID: "f7df4d1e-71e7-4c40-a3e4-27b484799d09"). InnerVolumeSpecName "kube-api-access-qj8fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.736510 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f7df4d1e-71e7-4c40-a3e4-27b484799d09" (UID: "f7df4d1e-71e7-4c40-a3e4-27b484799d09"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.736754 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-scripts" (OuterVolumeSpecName: "scripts") pod "f7df4d1e-71e7-4c40-a3e4-27b484799d09" (UID: "f7df4d1e-71e7-4c40-a3e4-27b484799d09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.792016 4733 scope.go:117] "RemoveContainer" containerID="a2f20b96c0e76358892a2e4c8789ed0cb1a66ac87763e230752a2979798603be" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.813282 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkmgs\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-kube-api-access-xkmgs\") pod \"d78c8335-c2e4-4c49-8b77-98fec3cab751\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.813397 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-combined-ca-bundle\") pod \"d78c8335-c2e4-4c49-8b77-98fec3cab751\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.813420 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-internal-tls-certs\") pod \"d78c8335-c2e4-4c49-8b77-98fec3cab751\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.813458 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-config-data\") pod \"d78c8335-c2e4-4c49-8b77-98fec3cab751\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.813515 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-run-httpd\") pod \"d78c8335-c2e4-4c49-8b77-98fec3cab751\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.813554 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-etc-swift\") pod \"d78c8335-c2e4-4c49-8b77-98fec3cab751\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.813592 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-public-tls-certs\") pod \"d78c8335-c2e4-4c49-8b77-98fec3cab751\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.813642 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-log-httpd\") pod \"d78c8335-c2e4-4c49-8b77-98fec3cab751\" (UID: \"d78c8335-c2e4-4c49-8b77-98fec3cab751\") " Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.814012 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.814023 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj8fc\" (UniqueName: \"kubernetes.io/projected/f7df4d1e-71e7-4c40-a3e4-27b484799d09-kube-api-access-qj8fc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.814033 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.814041 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7df4d1e-71e7-4c40-a3e4-27b484799d09-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.815582 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d78c8335-c2e4-4c49-8b77-98fec3cab751" (UID: "d78c8335-c2e4-4c49-8b77-98fec3cab751"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.818755 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d78c8335-c2e4-4c49-8b77-98fec3cab751" (UID: "d78c8335-c2e4-4c49-8b77-98fec3cab751"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.819660 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-kube-api-access-xkmgs" (OuterVolumeSpecName: "kube-api-access-xkmgs") pod "d78c8335-c2e4-4c49-8b77-98fec3cab751" (UID: "d78c8335-c2e4-4c49-8b77-98fec3cab751"). InnerVolumeSpecName "kube-api-access-xkmgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.819843 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d78c8335-c2e4-4c49-8b77-98fec3cab751" (UID: "d78c8335-c2e4-4c49-8b77-98fec3cab751"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.824476 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7df4d1e-71e7-4c40-a3e4-27b484799d09" (UID: "f7df4d1e-71e7-4c40-a3e4-27b484799d09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.872605 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.877454 4733 scope.go:117] "RemoveContainer" containerID="f5af192fa5fec93409d161a4e7f3e732ea9247d6b52157e69e7c2f283fe6f052" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.878343 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.890544 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.915881 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.915910 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkmgs\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-kube-api-access-xkmgs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.915919 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d78c8335-c2e4-4c49-8b77-98fec3cab751-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.915929 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.915938 4733 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d78c8335-c2e4-4c49-8b77-98fec3cab751-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:43 crc kubenswrapper[4733]: E1204 18:03:43.916004 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:43 crc kubenswrapper[4733]: E1204 18:03:43.916059 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data podName:5e605c62-64b4-4417-80bb-bc3387881f7a nodeName:}" failed. No retries permitted until 2025-12-04 18:03:47.916041372 +0000 UTC m=+1489.871402428 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data") pod "rabbitmq-cell1-server-0" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a") : configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.916391 4733 scope.go:117] "RemoveContainer" containerID="afe99239cae679354f70187888627dd752b6e0f56cc59225d53f0c36a6b659f8" Dec 04 18:03:43 crc kubenswrapper[4733]: I1204 18:03:43.941357 4733 scope.go:117] "RemoveContainer" containerID="c3b464a1016d052759179bfb0e7253393babdea9ae2efb95d1c7ade0f58ff289" Dec 04 18:03:43 crc kubenswrapper[4733]: E1204 18:03:43.944353 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7446c8ed_a97b_41a1_ae84_8f1eed8b203b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7446c8ed_a97b_41a1_ae84_8f1eed8b203b.slice/crio-e7892397ee71ab39db17b4f9ce076b0a9c65abc9b24ca5e76b41e706b70aa675\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae61093e_d566_48cb_acc2_dd88c4512b8e.slice/crio-162cddb3f0e641814c81d2b1c26c22e815c3678b6df38c946878e1635dfee126\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae61093e_d566_48cb_acc2_dd88c4512b8e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbae2e2ab_b6af_4d4d_8288_9f23050553a0.slice/crio-f70927bc9d53b516e4d204c03b0fb379b1c49f82a18c06f3a57df744e6f37932\": RecentStats: unable to find data in memory cache]" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.016694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5868\" (UniqueName: \"kubernetes.io/projected/4047083c-6418-4814-8b24-9407494d0303-kube-api-access-h5868\") pod \"4047083c-6418-4814-8b24-9407494d0303\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.017109 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-combined-ca-bundle\") pod \"4047083c-6418-4814-8b24-9407494d0303\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.017199 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-vencrypt-tls-certs\") pod \"4047083c-6418-4814-8b24-9407494d0303\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.017339 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-config-data\") pod \"4047083c-6418-4814-8b24-9407494d0303\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.017410 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-nova-novncproxy-tls-certs\") pod \"4047083c-6418-4814-8b24-9407494d0303\" (UID: \"4047083c-6418-4814-8b24-9407494d0303\") " Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.023600 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4047083c-6418-4814-8b24-9407494d0303-kube-api-access-h5868" (OuterVolumeSpecName: "kube-api-access-h5868") pod "4047083c-6418-4814-8b24-9407494d0303" (UID: "4047083c-6418-4814-8b24-9407494d0303"). InnerVolumeSpecName "kube-api-access-h5868". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.075528 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data" (OuterVolumeSpecName: "config-data") pod "f7df4d1e-71e7-4c40-a3e4-27b484799d09" (UID: "f7df4d1e-71e7-4c40-a3e4-27b484799d09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.119993 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.120314 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data podName:1f55e362-d59c-4269-92c3-d5ca014a2ef1 nodeName:}" failed. No retries permitted until 2025-12-04 18:03:48.120297845 +0000 UTC m=+1490.075658891 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data") pod "rabbitmq-server-0" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1") : configmap "rabbitmq-config-data" not found Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.120099 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5868\" (UniqueName: \"kubernetes.io/projected/4047083c-6418-4814-8b24-9407494d0303-kube-api-access-h5868\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.120400 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7df4d1e-71e7-4c40-a3e4-27b484799d09-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.204031 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.206258 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.206914 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.206957 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.208356 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.212580 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.215897 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.215950 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.286582 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d78c8335-c2e4-4c49-8b77-98fec3cab751" (UID: "d78c8335-c2e4-4c49-8b77-98fec3cab751"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.298636 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-config-data" (OuterVolumeSpecName: "config-data") pod "4047083c-6418-4814-8b24-9407494d0303" (UID: "4047083c-6418-4814-8b24-9407494d0303"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.323958 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.323980 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.352209 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18512067-9d64-4796-9bc1-a3d83ef5b13e" path="/var/lib/kubelet/pods/18512067-9d64-4796-9bc1-a3d83ef5b13e/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.352924 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" path="/var/lib/kubelet/pods/229fa5f0-e006-4c04-8476-23a7842d292b/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.354009 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e72addf-edb3-4704-be2e-206825774d97" path="/var/lib/kubelet/pods/4e72addf-edb3-4704-be2e-206825774d97/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.357143 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" path="/var/lib/kubelet/pods/7446c8ed-a97b-41a1-ae84-8f1eed8b203b/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.357768 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b19ede9-4e23-416d-bc7a-3674ed551c21" path="/var/lib/kubelet/pods/7b19ede9-4e23-416d-bc7a-3674ed551c21/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.358396 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" path="/var/lib/kubelet/pods/ae61093e-d566-48cb-acc2-dd88c4512b8e/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.359688 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" path="/var/lib/kubelet/pods/bae2e2ab-b6af-4d4d-8288-9f23050553a0/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.361218 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd664be9-e9ef-4734-8d88-b990f09de1f9" path="/var/lib/kubelet/pods/bd664be9-e9ef-4734-8d88-b990f09de1f9/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.365274 4733 generic.go:334] "Generic (PLEG): container finished" podID="5833da06-817f-4451-be68-9a9a94e5ef64" containerID="51718f0a98bc4f12648ae0c8e161095d9435571807a28e34ed23fb1ce1e33217" exitCode=0 Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.367278 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf38da41-b6c0-4551-9ce9-32146ebc2160" path="/var/lib/kubelet/pods/bf38da41-b6c0-4551-9ce9-32146ebc2160/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.370193 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d45f2b81-a9f9-47bc-8042-c344bffe4dac" path="/var/lib/kubelet/pods/d45f2b81-a9f9-47bc-8042-c344bffe4dac/volumes" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.377595 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-59945f4865-9zsr4" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.390060 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d78c8335-c2e4-4c49-8b77-98fec3cab751" (UID: "d78c8335-c2e4-4c49-8b77-98fec3cab751"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.395768 4733 generic.go:334] "Generic (PLEG): container finished" podID="32514f48-c09d-4bfc-a407-bc0b454ca946" containerID="1aa9e2ec45e1b882430e2811102f36451e65ac3c8ef29ffaad0b50cf3519c1a3" exitCode=0 Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.398696 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapi5c0f-account-delete-zk9rz" podStartSLOduration=4.398655977 podStartE2EDuration="4.398655977s" podCreationTimestamp="2025-12-04 18:03:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:03:44.388097087 +0000 UTC m=+1486.343458133" watchObservedRunningTime="2025-12-04 18:03:44.398655977 +0000 UTC m=+1486.354017023" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.400628 4733 generic.go:334] "Generic (PLEG): container finished" podID="4047083c-6418-4814-8b24-9407494d0303" containerID="55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1" exitCode=0 Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.400742 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.401060 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-config-data" (OuterVolumeSpecName: "config-data") pod "d78c8335-c2e4-4c49-8b77-98fec3cab751" (UID: "d78c8335-c2e4-4c49-8b77-98fec3cab751"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.417422 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4047083c-6418-4814-8b24-9407494d0303" (UID: "4047083c-6418-4814-8b24-9407494d0303"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.425286 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.425311 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.425320 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.459031 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d78c8335-c2e4-4c49-8b77-98fec3cab751" (UID: "d78c8335-c2e4-4c49-8b77-98fec3cab751"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.495785 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "4047083c-6418-4814-8b24-9407494d0303" (UID: "4047083c-6418-4814-8b24-9407494d0303"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.510970 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "4047083c-6418-4814-8b24-9407494d0303" (UID: "4047083c-6418-4814-8b24-9407494d0303"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.515711 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.517010 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.535097 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican2aa9-account-delete-7bq8x" podStartSLOduration=4.535078039 podStartE2EDuration="4.535078039s" podCreationTimestamp="2025-12-04 18:03:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:03:44.503417788 +0000 UTC m=+1486.458778834" watchObservedRunningTime="2025-12-04 18:03:44.535078039 +0000 UTC m=+1486.490439085" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.537822 4733 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.537851 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d78c8335-c2e4-4c49-8b77-98fec3cab751-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.537860 4733 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4047083c-6418-4814-8b24-9407494d0303-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.577513 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.608873 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placementee67-account-delete-sxlxx" podStartSLOduration=5.608855688 podStartE2EDuration="5.608855688s" podCreationTimestamp="2025-12-04 18:03:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:03:44.605642563 +0000 UTC m=+1486.561003609" watchObservedRunningTime="2025-12-04 18:03:44.608855688 +0000 UTC m=+1486.564216734" Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.652606 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:44 crc kubenswrapper[4733]: E1204 18:03:44.652680 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts podName:08cdf00b-cf60-4e3a-9283-833e3a8d95bf nodeName:}" failed. No retries permitted until 2025-12-04 18:03:45.152662191 +0000 UTC m=+1487.108023237 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts") pod "placementee67-account-delete-sxlxx" (UID: "08cdf00b-cf60-4e3a-9283-833e3a8d95bf") : configmap "openstack-scripts" not found Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.693839 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.774809 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.178:9292/healthcheck\": read tcp 10.217.0.2:57618->10.217.0.178:9292: read: connection reset by peer" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.774832 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.178:9292/healthcheck\": read tcp 10.217.0.2:57608->10.217.0.178:9292: read: connection reset by peer" Dec 04 18:03:44 crc kubenswrapper[4733]: I1204 18:03:44.824238 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.166:8776/healthcheck\": read tcp 10.217.0.2:53578->10.217.0.166:8776: read: connection reset by peer" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.115588 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5833da06-817f-4451-be68-9a9a94e5ef64","Type":"ContainerDied","Data":"51718f0a98bc4f12648ae0c8e161095d9435571807a28e34ed23fb1ce1e33217"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116000 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5833da06-817f-4451-be68-9a9a94e5ef64","Type":"ContainerDied","Data":"500dfcc1ae47b8d3bd343720e7a4af1ba091cdd3ba2dd568851fc0a4057d4c17"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116017 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="500dfcc1ae47b8d3bd343720e7a4af1ba091cdd3ba2dd568851fc0a4057d4c17" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116032 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116056 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116077 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi5c0f-account-delete-zk9rz" event={"ID":"faf216c0-9cda-4f76-a4df-8855352db991","Type":"ContainerStarted","Data":"b97820459144423fb9dffe72aae0571a0723b467879078ea36fb82747da5d291"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116097 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116120 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nkhjz"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116133 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-27f6z"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116143 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-64bd4cdd64-fftgx"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116158 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi5c0f-account-delete-zk9rz" event={"ID":"faf216c0-9cda-4f76-a4df-8855352db991","Type":"ContainerStarted","Data":"2a42bb3a9d25560a029b13a943a763cdfe7681f5633755f2a2552c368888e0c6"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116173 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59945f4865-9zsr4" event={"ID":"d78c8335-c2e4-4c49-8b77-98fec3cab751","Type":"ContainerDied","Data":"0cd6a42aad9bf69767b8f6ddd75025503a74ca5623b580aa89c5758a7b6ce2ec"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116190 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican2aa9-account-delete-7bq8x" event={"ID":"31d98e38-ee17-4974-94e7-6495564940be","Type":"ContainerStarted","Data":"c92d8d396adac2f52e75b63db8e67c9fd57abc43548e572275f0caf7dd9fb0ad"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116205 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-27f6z"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116222 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nkhjz"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116233 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116247 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell06ac2-account-delete-rs267" event={"ID":"32b94636-cf95-42f9-8589-e38b550053af","Type":"ContainerStarted","Data":"9ac864fb7141d34ebf8b982d955f2c128644efd586c33ac7b0cdf84e93109fc7"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116259 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-cron-29414521-gw72h"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116275 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-cron-29414521-gw72h"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116287 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-b92dw"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116300 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance776f-account-delete-696d2" event={"ID":"32514f48-c09d-4bfc-a407-bc0b454ca946","Type":"ContainerDied","Data":"1aa9e2ec45e1b882430e2811102f36451e65ac3c8ef29ffaad0b50cf3519c1a3"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116331 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance776f-account-delete-696d2" event={"ID":"32514f48-c09d-4bfc-a407-bc0b454ca946","Type":"ContainerStarted","Data":"c83599f693cb8a879141429622a802f59fe10e01e2c7c2563e439c9be03fc14a"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116346 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4047083c-6418-4814-8b24-9407494d0303","Type":"ContainerDied","Data":"55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116363 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-b92dw"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116378 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-df92-account-create-update-p899x"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116391 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-df92-account-create-update-p899x"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116403 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-zn7w9"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116414 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-zn7w9"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4047083c-6418-4814-8b24-9407494d0303","Type":"ContainerDied","Data":"ac2e572ab8366f12b677177b1a49a813aebef2886f0ed56f1e19766e5e6c543d"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116440 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron17fd-account-delete-pnhcm" event={"ID":"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc","Type":"ContainerStarted","Data":"10830977b29b2d58716fd705aedb7bc5d4ff1a61db024a027be6c97602a43587"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116453 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementee67-account-delete-sxlxx" event={"ID":"08cdf00b-cf60-4e3a-9283-833e3a8d95bf","Type":"ContainerStarted","Data":"ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116467 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ee67-account-create-update-56lqg"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116479 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementee67-account-delete-sxlxx" event={"ID":"08cdf00b-cf60-4e3a-9283-833e3a8d95bf","Type":"ContainerStarted","Data":"358da79aa4080332b38fbb6621db2dfd0b2f29bed19e6891cfa753d0cf58ca24"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116489 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementee67-account-delete-sxlxx"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116506 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ee67-account-create-update-56lqg"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.116958 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="ceilometer-central-agent" containerID="cri-o://43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.117171 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="658fbbbc-cb7d-4795-be06-76a46b589943" containerName="kube-state-metrics" containerID="cri-o://bd700eb5d3983184b0c921d502c52f9cada1865ba48dea23580767a8dbb28917" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.117309 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="3d28f771-2fba-42d9-8fc0-1107d17db527" containerName="memcached" containerID="cri-o://5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.117447 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-64bd4cdd64-fftgx" podUID="9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" containerName="keystone-api" containerID="cri-o://912469a5b6cd827fb51c283e6484beba8f59e32cbc52ed716d042adc91ed7c90" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.118026 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="proxy-httpd" containerID="cri-o://cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.118081 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="sg-core" containerID="cri-o://8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.118126 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="ceilometer-notification-agent" containerID="cri-o://aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.120329 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-t67cq"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.132947 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-t67cq"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.137829 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.166517 4733 scope.go:117] "RemoveContainer" containerID="11928a6e357a5e63f468c1daf6d2ee76e316437dd96f549ec172b0d953d18271" Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.172582 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.172640 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts podName:08cdf00b-cf60-4e3a-9283-833e3a8d95bf nodeName:}" failed. No retries permitted until 2025-12-04 18:03:46.172625768 +0000 UTC m=+1488.127986804 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts") pod "placementee67-account-delete-sxlxx" (UID: "08cdf00b-cf60-4e3a-9283-833e3a8d95bf") : configmap "openstack-scripts" not found Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.190436 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.205334 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance776f-account-delete-696d2"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.227174 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-776f-account-create-update-cfcmh"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.238516 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-776f-account-create-update-cfcmh"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.261149 4733 scope.go:117] "RemoveContainer" containerID="f130a979912194eb23dc33f38388f755ef2b29dc69498aa6ea5c028aff72030b" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.273763 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-default\") pod \"5833da06-817f-4451-be68-9a9a94e5ef64\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.273837 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-combined-ca-bundle\") pod \"5833da06-817f-4451-be68-9a9a94e5ef64\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.273948 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-galera-tls-certs\") pod \"5833da06-817f-4451-be68-9a9a94e5ef64\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.274007 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-kolla-config\") pod \"5833da06-817f-4451-be68-9a9a94e5ef64\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.274027 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-generated\") pod \"5833da06-817f-4451-be68-9a9a94e5ef64\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.274089 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5833da06-817f-4451-be68-9a9a94e5ef64\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.274114 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpf66\" (UniqueName: \"kubernetes.io/projected/5833da06-817f-4451-be68-9a9a94e5ef64-kube-api-access-bpf66\") pod \"5833da06-817f-4451-be68-9a9a94e5ef64\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.274134 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-operator-scripts\") pod \"5833da06-817f-4451-be68-9a9a94e5ef64\" (UID: \"5833da06-817f-4451-be68-9a9a94e5ef64\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.275301 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5833da06-817f-4451-be68-9a9a94e5ef64" (UID: "5833da06-817f-4451-be68-9a9a94e5ef64"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.275890 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "5833da06-817f-4451-be68-9a9a94e5ef64" (UID: "5833da06-817f-4451-be68-9a9a94e5ef64"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.293311 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "5833da06-817f-4451-be68-9a9a94e5ef64" (UID: "5833da06-817f-4451-be68-9a9a94e5ef64"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.303132 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "5833da06-817f-4451-be68-9a9a94e5ef64" (UID: "5833da06-817f-4451-be68-9a9a94e5ef64"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.303192 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.315178 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.323279 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-q4prp"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.328749 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5833da06-817f-4451-be68-9a9a94e5ef64-kube-api-access-bpf66" (OuterVolumeSpecName: "kube-api-access-bpf66") pod "5833da06-817f-4451-be68-9a9a94e5ef64" (UID: "5833da06-817f-4451-be68-9a9a94e5ef64"). InnerVolumeSpecName "kube-api-access-bpf66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.331641 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-q4prp"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.336994 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5833da06-817f-4451-be68-9a9a94e5ef64" (UID: "5833da06-817f-4451-be68-9a9a94e5ef64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.337748 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "mysql-db") pod "5833da06-817f-4451-be68-9a9a94e5ef64" (UID: "5833da06-817f-4451-be68-9a9a94e5ef64"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.347062 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-59945f4865-9zsr4"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.351281 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-59945f4865-9zsr4"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.365387 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-7119-account-create-update-k2cvp"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.370236 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "5833da06-817f-4451-be68-9a9a94e5ef64" (UID: "5833da06-817f-4451-be68-9a9a94e5ef64"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.372990 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder7119-account-delete-kzkjb"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.375328 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-config-data\") pod \"f575fafd-64f0-4401-82f1-f4a6a33b132e\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.375390 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-scripts\") pod \"f575fafd-64f0-4401-82f1-f4a6a33b132e\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.375495 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msvfc\" (UniqueName: \"kubernetes.io/projected/f575fafd-64f0-4401-82f1-f4a6a33b132e-kube-api-access-msvfc\") pod \"f575fafd-64f0-4401-82f1-f4a6a33b132e\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.375591 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-internal-tls-certs\") pod \"f575fafd-64f0-4401-82f1-f4a6a33b132e\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.375619 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-combined-ca-bundle\") pod \"f575fafd-64f0-4401-82f1-f4a6a33b132e\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.375652 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-public-tls-certs\") pod \"f575fafd-64f0-4401-82f1-f4a6a33b132e\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.375672 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f575fafd-64f0-4401-82f1-f4a6a33b132e-logs\") pod \"f575fafd-64f0-4401-82f1-f4a6a33b132e\" (UID: \"f575fafd-64f0-4401-82f1-f4a6a33b132e\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.376044 4733 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.376059 4733 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.376070 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.376089 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.376098 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpf66\" (UniqueName: \"kubernetes.io/projected/5833da06-817f-4451-be68-9a9a94e5ef64-kube-api-access-bpf66\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.376108 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.376116 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5833da06-817f-4451-be68-9a9a94e5ef64-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.376126 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5833da06-817f-4451-be68-9a9a94e5ef64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.379616 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f575fafd-64f0-4401-82f1-f4a6a33b132e-logs" (OuterVolumeSpecName: "logs") pod "f575fafd-64f0-4401-82f1-f4a6a33b132e" (UID: "f575fafd-64f0-4401-82f1-f4a6a33b132e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.379961 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-scripts" (OuterVolumeSpecName: "scripts") pod "f575fafd-64f0-4401-82f1-f4a6a33b132e" (UID: "f575fafd-64f0-4401-82f1-f4a6a33b132e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.380274 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-7119-account-create-update-k2cvp"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.381485 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f575fafd-64f0-4401-82f1-f4a6a33b132e-kube-api-access-msvfc" (OuterVolumeSpecName: "kube-api-access-msvfc") pod "f575fafd-64f0-4401-82f1-f4a6a33b132e" (UID: "f575fafd-64f0-4401-82f1-f4a6a33b132e"). InnerVolumeSpecName "kube-api-access-msvfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.387826 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.394390 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.453014 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-76hpp"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.456308 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.461848 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-76hpp"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.469930 4733 scope.go:117] "RemoveContainer" containerID="7a464410f79ac8c4ad9cc078a988fa37fc20c807e7742c3550f9f6de916082bb" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.471925 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": read tcp 10.217.0.2:50676->10.217.0.207:8775: read: connection reset by peer" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.477554 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.477586 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f575fafd-64f0-4401-82f1-f4a6a33b132e-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.477598 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.477609 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msvfc\" (UniqueName: \"kubernetes.io/projected/f575fafd-64f0-4401-82f1-f4a6a33b132e-kube-api-access-msvfc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.485787 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-2aa9-account-create-update-dd8zp"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.485967 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": read tcp 10.217.0.2:50664->10.217.0.207:8775: read: connection reset by peer" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.493144 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-2aa9-account-create-update-dd8zp"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.504939 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican2aa9-account-delete-7bq8x"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.509468 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7f757c77dd-bzp5x" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:45484->10.217.0.165:9311: read: connection reset by peer" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.509538 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7f757c77dd-bzp5x" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:45480->10.217.0.165:9311: read: connection reset by peer" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.514397 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerName="galera" containerID="cri-o://e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.522046 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f575fafd-64f0-4401-82f1-f4a6a33b132e" (UID: "f575fafd-64f0-4401-82f1-f4a6a33b132e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.525442 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-config-data" (OuterVolumeSpecName: "config-data") pod "f575fafd-64f0-4401-82f1-f4a6a33b132e" (UID: "f575fafd-64f0-4401-82f1-f4a6a33b132e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.536994 4733 generic.go:334] "Generic (PLEG): container finished" podID="8b77d7a1-e0fe-466e-9629-ac6b5bab61dc" containerID="10830977b29b2d58716fd705aedb7bc5d4ff1a61db024a027be6c97602a43587" exitCode=0 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.537063 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron17fd-account-delete-pnhcm" event={"ID":"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc","Type":"ContainerDied","Data":"10830977b29b2d58716fd705aedb7bc5d4ff1a61db024a027be6c97602a43587"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.541402 4733 generic.go:334] "Generic (PLEG): container finished" podID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerID="5cd076e1396593ece03d523ffdfe9f53cc72e35677e8aab21b5111c2bcce8fa1" exitCode=0 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.541468 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12b58db1-9574-4081-a3c1-2853f003fa8f","Type":"ContainerDied","Data":"5cd076e1396593ece03d523ffdfe9f53cc72e35677e8aab21b5111c2bcce8fa1"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.541497 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12b58db1-9574-4081-a3c1-2853f003fa8f","Type":"ContainerDied","Data":"3545c73604d8f3cb46724ea7074e847de530750900a13c82a111f09584041895"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.541510 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3545c73604d8f3cb46724ea7074e847de530750900a13c82a111f09584041895" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.544385 4733 generic.go:334] "Generic (PLEG): container finished" podID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerID="6560f5a95ae3f9b652dd82bb257bd795655e85bb32c02059e6da7c32d0b36e95" exitCode=0 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.544448 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d","Type":"ContainerDied","Data":"6560f5a95ae3f9b652dd82bb257bd795655e85bb32c02059e6da7c32d0b36e95"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.547246 4733 generic.go:334] "Generic (PLEG): container finished" podID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerID="8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7" exitCode=2 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.547306 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerDied","Data":"8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.557271 4733 generic.go:334] "Generic (PLEG): container finished" podID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerID="cbcf529109c0c4af1bdf871c346e94fb3c2778251cd7b3b1d1b5864a31229cb3" exitCode=0 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.557340 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c8b0056-e347-4f24-aeea-20680c33bf06","Type":"ContainerDied","Data":"cbcf529109c0c4af1bdf871c346e94fb3c2778251cd7b3b1d1b5864a31229cb3"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.572055 4733 generic.go:334] "Generic (PLEG): container finished" podID="31d98e38-ee17-4974-94e7-6495564940be" containerID="c92d8d396adac2f52e75b63db8e67c9fd57abc43548e572275f0caf7dd9fb0ad" exitCode=0 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.572932 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican2aa9-account-delete-7bq8x" event={"ID":"31d98e38-ee17-4974-94e7-6495564940be","Type":"ContainerDied","Data":"c92d8d396adac2f52e75b63db8e67c9fd57abc43548e572275f0caf7dd9fb0ad"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.582820 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.582859 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.587559 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f575fafd-64f0-4401-82f1-f4a6a33b132e" (UID: "f575fafd-64f0-4401-82f1-f4a6a33b132e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.589508 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell06ac2-account-delete-rs267" event={"ID":"32b94636-cf95-42f9-8589-e38b550053af","Type":"ContainerStarted","Data":"ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.589979 4733 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell06ac2-account-delete-rs267" secret="" err="secret \"galera-openstack-dockercfg-c4nlc\" not found" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.596469 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.597774 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-5vq76"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.611625 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-5vq76"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.613294 4733 scope.go:117] "RemoveContainer" containerID="55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.616513 4733 generic.go:334] "Generic (PLEG): container finished" podID="658fbbbc-cb7d-4795-be06-76a46b589943" containerID="bd700eb5d3983184b0c921d502c52f9cada1865ba48dea23580767a8dbb28917" exitCode=2 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.616664 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"658fbbbc-cb7d-4795-be06-76a46b589943","Type":"ContainerDied","Data":"bd700eb5d3983184b0c921d502c52f9cada1865ba48dea23580767a8dbb28917"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.617219 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.619683 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder7119-account-delete-kzkjb" event={"ID":"3f1c1320-5093-464a-a9a0-8a4cb668d70f","Type":"ContainerStarted","Data":"1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.619855 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f575fafd-64f0-4401-82f1-f4a6a33b132e" (UID: "f575fafd-64f0-4401-82f1-f4a6a33b132e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.620318 4733 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder7119-account-delete-kzkjb" secret="" err="secret \"galera-openstack-dockercfg-c4nlc\" not found" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.622347 4733 generic.go:334] "Generic (PLEG): container finished" podID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerID="d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8" exitCode=0 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.622386 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6df7889d-qkp4j" event={"ID":"f575fafd-64f0-4401-82f1-f4a6a33b132e","Type":"ContainerDied","Data":"d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.622405 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d6df7889d-qkp4j" event={"ID":"f575fafd-64f0-4401-82f1-f4a6a33b132e","Type":"ContainerDied","Data":"c9ac1c8691d89603fcbd0ce69f6273af7294c2c5fb8416393ad0228a2ed56f2b"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.622450 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d6df7889d-qkp4j" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.630173 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-17fd-account-create-update-ct9z4"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.637121 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron17fd-account-delete-pnhcm"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.651304 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-17fd-account-create-update-ct9z4"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.656564 4733 generic.go:334] "Generic (PLEG): container finished" podID="faf216c0-9cda-4f76-a4df-8855352db991" containerID="b97820459144423fb9dffe72aae0571a0723b467879078ea36fb82747da5d291" exitCode=0 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.656775 4733 scope.go:117] "RemoveContainer" containerID="55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.656910 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi5c0f-account-delete-zk9rz" event={"ID":"faf216c0-9cda-4f76-a4df-8855352db991","Type":"ContainerDied","Data":"b97820459144423fb9dffe72aae0571a0723b467879078ea36fb82747da5d291"} Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.657015 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placementee67-account-delete-sxlxx" podUID="08cdf00b-cf60-4e3a-9283-833e3a8d95bf" containerName="mariadb-account-delete" containerID="cri-o://ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a" gracePeriod=30 Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.657263 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.661560 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell06ac2-account-delete-rs267" podStartSLOduration=4.661540879 podStartE2EDuration="4.661540879s" podCreationTimestamp="2025-12-04 18:03:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:03:45.654742639 +0000 UTC m=+1487.610103675" watchObservedRunningTime="2025-12-04 18:03:45.661540879 +0000 UTC m=+1487.616901925" Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.662219 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1\": container with ID starting with 55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1 not found: ID does not exist" containerID="55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.662267 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1"} err="failed to get container status \"55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1\": rpc error: code = NotFound desc = could not find container \"55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1\": container with ID starting with 55689e963ef8e25c2f4bd316d8235eac66d987ac3825cf8882cc134a4824acf1 not found: ID does not exist" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.662297 4733 scope.go:117] "RemoveContainer" containerID="d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.692848 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12b58db1-9574-4081-a3c1-2853f003fa8f-logs\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.692904 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data-custom\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.692948 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-combined-ca-bundle\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.693428 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12b58db1-9574-4081-a3c1-2853f003fa8f-logs" (OuterVolumeSpecName: "logs") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.693526 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.693619 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dtf5\" (UniqueName: \"kubernetes.io/projected/12b58db1-9574-4081-a3c1-2853f003fa8f-kube-api-access-8dtf5\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.693649 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-scripts\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.693707 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12b58db1-9574-4081-a3c1-2853f003fa8f-etc-machine-id\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.696268 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12b58db1-9574-4081-a3c1-2853f003fa8f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.697173 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.698056 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-scripts" (OuterVolumeSpecName: "scripts") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.702524 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-internal-tls-certs\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.702634 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-public-tls-certs\") pod \"12b58db1-9574-4081-a3c1-2853f003fa8f\" (UID: \"12b58db1-9574-4081-a3c1-2853f003fa8f\") " Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.703224 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.703309 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts podName:32b94636-cf95-42f9-8589-e38b550053af nodeName:}" failed. No retries permitted until 2025-12-04 18:03:46.203288068 +0000 UTC m=+1488.158649114 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts") pod "novacell06ac2-account-delete-rs267" (UID: "32b94636-cf95-42f9-8589-e38b550053af") : configmap "openstack-scripts" not found Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.715271 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.715414 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f575fafd-64f0-4401-82f1-f4a6a33b132e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.715431 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.715444 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12b58db1-9574-4081-a3c1-2853f003fa8f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.715456 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12b58db1-9574-4081-a3c1-2853f003fa8f-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.715467 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.716158 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12b58db1-9574-4081-a3c1-2853f003fa8f-kube-api-access-8dtf5" (OuterVolumeSpecName: "kube-api-access-8dtf5") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "kube-api-access-8dtf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.729111 4733 scope.go:117] "RemoveContainer" containerID="606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.778711 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder7119-account-delete-kzkjb" podStartSLOduration=5.77869498 podStartE2EDuration="5.77869498s" podCreationTimestamp="2025-12-04 18:03:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 18:03:45.722858207 +0000 UTC m=+1487.678219253" watchObservedRunningTime="2025-12-04 18:03:45.77869498 +0000 UTC m=+1487.734056026" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.781652 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.788822 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d6df7889d-qkp4j"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.789559 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.796788 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-d6df7889d-qkp4j"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.800929 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data" (OuterVolumeSpecName: "config-data") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.804409 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "12b58db1-9574-4081-a3c1-2853f003fa8f" (UID: "12b58db1-9574-4081-a3c1-2853f003fa8f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.809102 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.814676 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817117 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-public-tls-certs\") pod \"2c8b0056-e347-4f24-aeea-20680c33bf06\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817156 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-config-data\") pod \"2c8b0056-e347-4f24-aeea-20680c33bf06\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817200 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-scripts\") pod \"2c8b0056-e347-4f24-aeea-20680c33bf06\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817282 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-combined-ca-bundle\") pod \"2c8b0056-e347-4f24-aeea-20680c33bf06\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817309 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hhhk\" (UniqueName: \"kubernetes.io/projected/2c8b0056-e347-4f24-aeea-20680c33bf06-kube-api-access-7hhhk\") pod \"2c8b0056-e347-4f24-aeea-20680c33bf06\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817358 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-logs\") pod \"2c8b0056-e347-4f24-aeea-20680c33bf06\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817412 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2c8b0056-e347-4f24-aeea-20680c33bf06\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817473 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-httpd-run\") pod \"2c8b0056-e347-4f24-aeea-20680c33bf06\" (UID: \"2c8b0056-e347-4f24-aeea-20680c33bf06\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817868 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817884 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dtf5\" (UniqueName: \"kubernetes.io/projected/12b58db1-9574-4081-a3c1-2853f003fa8f-kube-api-access-8dtf5\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817894 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817903 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.817911 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b58db1-9574-4081-a3c1-2853f003fa8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.821125 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.823462 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c8b0056-e347-4f24-aeea-20680c33bf06-kube-api-access-7hhhk" (OuterVolumeSpecName: "kube-api-access-7hhhk") pod "2c8b0056-e347-4f24-aeea-20680c33bf06" (UID: "2c8b0056-e347-4f24-aeea-20680c33bf06"). InnerVolumeSpecName "kube-api-access-7hhhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.824046 4733 scope.go:117] "RemoveContainer" containerID="d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.824374 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-scripts" (OuterVolumeSpecName: "scripts") pod "2c8b0056-e347-4f24-aeea-20680c33bf06" (UID: "2c8b0056-e347-4f24-aeea-20680c33bf06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.824410 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2c8b0056-e347-4f24-aeea-20680c33bf06" (UID: "2c8b0056-e347-4f24-aeea-20680c33bf06"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.824447 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.824518 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8\": container with ID starting with d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8 not found: ID does not exist" containerID="d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.824545 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8"} err="failed to get container status \"d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8\": rpc error: code = NotFound desc = could not find container \"d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8\": container with ID starting with d9729a8279a25f8543117eca27b8da8fddae80ef7b81b4f86a36adf2de5efdd8 not found: ID does not exist" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.824578 4733 scope.go:117] "RemoveContainer" containerID="606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2" Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.824529 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts podName:3f1c1320-5093-464a-a9a0-8a4cb668d70f nodeName:}" failed. No retries permitted until 2025-12-04 18:03:46.324511867 +0000 UTC m=+1488.279872913 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts") pod "cinder7119-account-delete-kzkjb" (UID: "3f1c1320-5093-464a-a9a0-8a4cb668d70f") : configmap "openstack-scripts" not found Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.824817 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-logs" (OuterVolumeSpecName: "logs") pod "2c8b0056-e347-4f24-aeea-20680c33bf06" (UID: "2c8b0056-e347-4f24-aeea-20680c33bf06"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.824855 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2\": container with ID starting with 606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2 not found: ID does not exist" containerID="606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.824884 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2"} err="failed to get container status \"606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2\": rpc error: code = NotFound desc = could not find container \"606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2\": container with ID starting with 606b6c780e5c6e2d4acb2a66a0dfa3ab1c40dbc514201e9046a42a373b70f4d2 not found: ID does not exist" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.836552 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "2c8b0056-e347-4f24-aeea-20680c33bf06" (UID: "2c8b0056-e347-4f24-aeea-20680c33bf06"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.863456 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c8b0056-e347-4f24-aeea-20680c33bf06" (UID: "2c8b0056-e347-4f24-aeea-20680c33bf06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.866781 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.905169 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.906340 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2c8b0056-e347-4f24-aeea-20680c33bf06" (UID: "2c8b0056-e347-4f24-aeea-20680c33bf06"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.906928 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.908149 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 04 18:03:45 crc kubenswrapper[4733]: E1204 18:03:45.908230 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerName="galera" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.910845 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-config-data" (OuterVolumeSpecName: "config-data") pod "2c8b0056-e347-4f24-aeea-20680c33bf06" (UID: "2c8b0056-e347-4f24-aeea-20680c33bf06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.920979 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-certs\") pod \"658fbbbc-cb7d-4795-be06-76a46b589943\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921074 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-combined-ca-bundle\") pod \"658fbbbc-cb7d-4795-be06-76a46b589943\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921133 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-config\") pod \"658fbbbc-cb7d-4795-be06-76a46b589943\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921172 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjmk2\" (UniqueName: \"kubernetes.io/projected/658fbbbc-cb7d-4795-be06-76a46b589943-kube-api-access-sjmk2\") pod \"658fbbbc-cb7d-4795-be06-76a46b589943\" (UID: \"658fbbbc-cb7d-4795-be06-76a46b589943\") " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921528 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921551 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921563 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921574 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921584 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8b0056-e347-4f24-aeea-20680c33bf06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921594 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hhhk\" (UniqueName: \"kubernetes.io/projected/2c8b0056-e347-4f24-aeea-20680c33bf06-kube-api-access-7hhhk\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921604 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c8b0056-e347-4f24-aeea-20680c33bf06-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.921624 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.925660 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658fbbbc-cb7d-4795-be06-76a46b589943-kube-api-access-sjmk2" (OuterVolumeSpecName: "kube-api-access-sjmk2") pod "658fbbbc-cb7d-4795-be06-76a46b589943" (UID: "658fbbbc-cb7d-4795-be06-76a46b589943"). InnerVolumeSpecName "kube-api-access-sjmk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.942034 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.951218 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "658fbbbc-cb7d-4795-be06-76a46b589943" (UID: "658fbbbc-cb7d-4795-be06-76a46b589943"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.957187 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "658fbbbc-cb7d-4795-be06-76a46b589943" (UID: "658fbbbc-cb7d-4795-be06-76a46b589943"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.957341 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:03:45 crc kubenswrapper[4733]: I1204 18:03:45.980955 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "658fbbbc-cb7d-4795-be06-76a46b589943" (UID: "658fbbbc-cb7d-4795-be06-76a46b589943"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.023370 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-httpd-run\") pod \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.023459 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-logs\") pod \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.023521 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-internal-tls-certs\") pod \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.023546 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-config-data\") pod \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.023571 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-combined-ca-bundle\") pod \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.023655 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.023768 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-scripts\") pod \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.023839 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9wpc\" (UniqueName: \"kubernetes.io/projected/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-kube-api-access-d9wpc\") pod \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\" (UID: \"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.024247 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-logs" (OuterVolumeSpecName: "logs") pod "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" (UID: "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.024431 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.024451 4733 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.024465 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.024475 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.024484 4733 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/658fbbbc-cb7d-4795-be06-76a46b589943-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.024494 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjmk2\" (UniqueName: \"kubernetes.io/projected/658fbbbc-cb7d-4795-be06-76a46b589943-kube-api-access-sjmk2\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.024923 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" (UID: "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.036970 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" (UID: "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.036990 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-scripts" (OuterVolumeSpecName: "scripts") pod "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" (UID: "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.052160 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-kube-api-access-d9wpc" (OuterVolumeSpecName: "kube-api-access-d9wpc") pod "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" (UID: "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d"). InnerVolumeSpecName "kube-api-access-d9wpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.053159 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-vk5ks"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.061436 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-vk5ks"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.061490 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" (UID: "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.079616 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5c0f-account-create-update-2gwwc"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.080224 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" (UID: "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.090120 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi5c0f-account-delete-zk9rz"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.106776 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5c0f-account-create-update-2gwwc"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.115148 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-config-data" (OuterVolumeSpecName: "config-data") pod "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" (UID: "98545dd5-cfa2-4c6c-97e9-6985a1d5a60d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.125808 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd72019f-8a36-43af-8d73-7c3e804c1baa-logs\") pod \"dd72019f-8a36-43af-8d73-7c3e804c1baa\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126020 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-config-data\") pod \"dd72019f-8a36-43af-8d73-7c3e804c1baa\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126055 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-combined-ca-bundle\") pod \"dd72019f-8a36-43af-8d73-7c3e804c1baa\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126083 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm4dx\" (UniqueName: \"kubernetes.io/projected/dd72019f-8a36-43af-8d73-7c3e804c1baa-kube-api-access-vm4dx\") pod \"dd72019f-8a36-43af-8d73-7c3e804c1baa\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126121 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-nova-metadata-tls-certs\") pod \"dd72019f-8a36-43af-8d73-7c3e804c1baa\" (UID: \"dd72019f-8a36-43af-8d73-7c3e804c1baa\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126491 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126505 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126514 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9wpc\" (UniqueName: \"kubernetes.io/projected/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-kube-api-access-d9wpc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126523 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126532 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126542 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.126550 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.131345 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd72019f-8a36-43af-8d73-7c3e804c1baa-logs" (OuterVolumeSpecName: "logs") pod "dd72019f-8a36-43af-8d73-7c3e804c1baa" (UID: "dd72019f-8a36-43af-8d73-7c3e804c1baa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.132633 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd72019f-8a36-43af-8d73-7c3e804c1baa-kube-api-access-vm4dx" (OuterVolumeSpecName: "kube-api-access-vm4dx") pod "dd72019f-8a36-43af-8d73-7c3e804c1baa" (UID: "dd72019f-8a36-43af-8d73-7c3e804c1baa"). InnerVolumeSpecName "kube-api-access-vm4dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.149785 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.176777 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd72019f-8a36-43af-8d73-7c3e804c1baa" (UID: "dd72019f-8a36-43af-8d73-7c3e804c1baa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.217696 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-r28fb"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.230826 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.230862 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.230885 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vm4dx\" (UniqueName: \"kubernetes.io/projected/dd72019f-8a36-43af-8d73-7c3e804c1baa-kube-api-access-vm4dx\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.230894 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd72019f-8a36-43af-8d73-7c3e804c1baa-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.230952 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.231004 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts podName:32b94636-cf95-42f9-8589-e38b550053af nodeName:}" failed. No retries permitted until 2025-12-04 18:03:47.23098857 +0000 UTC m=+1489.186349616 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts") pod "novacell06ac2-account-delete-rs267" (UID: "32b94636-cf95-42f9-8589-e38b550053af") : configmap "openstack-scripts" not found Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.231171 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.231219 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts podName:08cdf00b-cf60-4e3a-9283-833e3a8d95bf nodeName:}" failed. No retries permitted until 2025-12-04 18:03:48.231193776 +0000 UTC m=+1490.186554822 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts") pod "placementee67-account-delete-sxlxx" (UID: "08cdf00b-cf60-4e3a-9283-833e3a8d95bf") : configmap "openstack-scripts" not found Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.241078 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-r28fb"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.246905 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-config-data" (OuterVolumeSpecName: "config-data") pod "dd72019f-8a36-43af-8d73-7c3e804c1baa" (UID: "dd72019f-8a36-43af-8d73-7c3e804c1baa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.265429 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "dd72019f-8a36-43af-8d73-7c3e804c1baa" (UID: "dd72019f-8a36-43af-8d73-7c3e804c1baa"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.267547 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-6ac2-account-create-update-l2v8g"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.277512 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell06ac2-account-delete-rs267"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.283967 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-6ac2-account-create-update-l2v8g"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.332543 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.332575 4733 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd72019f-8a36-43af-8d73-7c3e804c1baa-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.332653 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.332708 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts podName:3f1c1320-5093-464a-a9a0-8a4cb668d70f nodeName:}" failed. No retries permitted until 2025-12-04 18:03:47.33269178 +0000 UTC m=+1489.288052826 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts") pod "cinder7119-account-delete-kzkjb" (UID: "3f1c1320-5093-464a-a9a0-8a4cb668d70f") : configmap "openstack-scripts" not found Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.349672 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6" path="/var/lib/kubelet/pods/0e9fda50-9c28-4c0d-8bdd-b4caadeaa7b6/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.350388 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16f02d93-bf6f-4123-8781-2238ffbac212" path="/var/lib/kubelet/pods/16f02d93-bf6f-4123-8781-2238ffbac212/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.351129 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b35785d-5a00-4554-b0ac-5585ca1c4816" path="/var/lib/kubelet/pods/1b35785d-5a00-4554-b0ac-5585ca1c4816/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.351817 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fca0e02-094d-4588-9efa-09181761155a" path="/var/lib/kubelet/pods/2fca0e02-094d-4588-9efa-09181761155a/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.353160 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37aa673f-32a7-43ce-b1b2-a7f02f46c485" path="/var/lib/kubelet/pods/37aa673f-32a7-43ce-b1b2-a7f02f46c485/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.353831 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4047083c-6418-4814-8b24-9407494d0303" path="/var/lib/kubelet/pods/4047083c-6418-4814-8b24-9407494d0303/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.354441 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40ea0fb9-d583-4ea8-8bb4-691e0f8606d3" path="/var/lib/kubelet/pods/40ea0fb9-d583-4ea8-8bb4-691e0f8606d3/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.355750 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41470d79-78b1-44bb-a10f-9f169d2d9e17" path="/var/lib/kubelet/pods/41470d79-78b1-44bb-a10f-9f169d2d9e17/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.356389 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50ad724d-2883-415e-97f4-636ee63d864c" path="/var/lib/kubelet/pods/50ad724d-2883-415e-97f4-636ee63d864c/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.357142 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" path="/var/lib/kubelet/pods/5833da06-817f-4451-be68-9a9a94e5ef64/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.358416 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b8dc93-1d10-4355-9d7f-ce423a3c11bc" path="/var/lib/kubelet/pods/62b8dc93-1d10-4355-9d7f-ce423a3c11bc/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.359081 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a32f53d-9569-405a-9060-80e298dfc6ce" path="/var/lib/kubelet/pods/6a32f53d-9569-405a-9060-80e298dfc6ce/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.359707 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75b050ca-b135-4564-be78-fd2a4efefdbe" path="/var/lib/kubelet/pods/75b050ca-b135-4564-be78-fd2a4efefdbe/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.360951 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95c361e7-6b27-4624-8a71-bccd6e6c0cd0" path="/var/lib/kubelet/pods/95c361e7-6b27-4624-8a71-bccd6e6c0cd0/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.361682 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bbdf249-90f9-4685-9418-1ecd4729ac09" path="/var/lib/kubelet/pods/9bbdf249-90f9-4685-9418-1ecd4729ac09/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.362555 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6b87ca2-4352-4387-8e5d-6070a278c63d" path="/var/lib/kubelet/pods/a6b87ca2-4352-4387-8e5d-6070a278c63d/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.363288 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b41b1c63-78a9-416c-9358-9a436fae72da" path="/var/lib/kubelet/pods/b41b1c63-78a9-416c-9358-9a436fae72da/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.364711 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c30a71d3-ad60-465b-88f2-e642c701f598" path="/var/lib/kubelet/pods/c30a71d3-ad60-465b-88f2-e642c701f598/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.365426 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" path="/var/lib/kubelet/pods/d78c8335-c2e4-4c49-8b77-98fec3cab751/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.366152 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7fdf3ca-8982-4da7-8b15-1647ddb2fef3" path="/var/lib/kubelet/pods/d7fdf3ca-8982-4da7-8b15-1647ddb2fef3/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.367371 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8436db7-0c0c-4c64-aea2-2266437d9d44" path="/var/lib/kubelet/pods/e8436db7-0c0c-4c64-aea2-2266437d9d44/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.368137 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3" path="/var/lib/kubelet/pods/ea7a064e-d1c8-4fd6-8755-f2e0b0e999e3/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.368851 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" path="/var/lib/kubelet/pods/f575fafd-64f0-4401-82f1-f4a6a33b132e/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.370144 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" path="/var/lib/kubelet/pods/f7df4d1e-71e7-4c40-a3e4-27b484799d09/volumes" Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.394575 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.396566 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.400290 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.400337 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="839d7537-ff12-4ee6-b36c-b192f130a6e4" containerName="nova-cell1-conductor-conductor" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.403439 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.426682 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.450150 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.456300 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.522658 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.204:3000/\": dial tcp 10.217.0.204:3000: connect: connection refused" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.536348 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-operator-scripts\") pod \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\" (UID: \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.536528 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlgtv\" (UniqueName: \"kubernetes.io/projected/32514f48-c09d-4bfc-a407-bc0b454ca946-kube-api-access-jlgtv\") pod \"32514f48-c09d-4bfc-a407-bc0b454ca946\" (UID: \"32514f48-c09d-4bfc-a407-bc0b454ca946\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.536693 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k8t2\" (UniqueName: \"kubernetes.io/projected/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-kube-api-access-5k8t2\") pod \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\" (UID: \"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.537463 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b77d7a1-e0fe-466e-9629-ac6b5bab61dc" (UID: "8b77d7a1-e0fe-466e-9629-ac6b5bab61dc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.537571 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32514f48-c09d-4bfc-a407-bc0b454ca946-operator-scripts\") pod \"32514f48-c09d-4bfc-a407-bc0b454ca946\" (UID: \"32514f48-c09d-4bfc-a407-bc0b454ca946\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.538407 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.538540 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32514f48-c09d-4bfc-a407-bc0b454ca946-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32514f48-c09d-4bfc-a407-bc0b454ca946" (UID: "32514f48-c09d-4bfc-a407-bc0b454ca946"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.542927 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-kube-api-access-5k8t2" (OuterVolumeSpecName: "kube-api-access-5k8t2") pod "8b77d7a1-e0fe-466e-9629-ac6b5bab61dc" (UID: "8b77d7a1-e0fe-466e-9629-ac6b5bab61dc"). InnerVolumeSpecName "kube-api-access-5k8t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.545976 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32514f48-c09d-4bfc-a407-bc0b454ca946-kube-api-access-jlgtv" (OuterVolumeSpecName: "kube-api-access-jlgtv") pod "32514f48-c09d-4bfc-a407-bc0b454ca946" (UID: "32514f48-c09d-4bfc-a407-bc0b454ca946"). InnerVolumeSpecName "kube-api-access-jlgtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.626216 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.639891 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52st8\" (UniqueName: \"kubernetes.io/projected/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-kube-api-access-52st8\") pod \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.639951 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f707420-7a54-4f2c-ac75-38e09d8dfb61-logs\") pod \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.639978 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-public-tls-certs\") pod \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640049 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-combined-ca-bundle\") pod \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640115 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-combined-ca-bundle\") pod \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640139 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data\") pod \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640168 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-public-tls-certs\") pod \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640212 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-internal-tls-certs\") pod \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640237 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn5tm\" (UniqueName: \"kubernetes.io/projected/2f707420-7a54-4f2c-ac75-38e09d8dfb61-kube-api-access-xn5tm\") pod \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640306 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-internal-tls-certs\") pod \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640361 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data-custom\") pod \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640384 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-config-data\") pod \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\" (UID: \"2f707420-7a54-4f2c-ac75-38e09d8dfb61\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640420 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-logs\") pod \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\" (UID: \"3c5f9b26-6c76-4af7-a811-d7d763f74ed6\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640899 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlgtv\" (UniqueName: \"kubernetes.io/projected/32514f48-c09d-4bfc-a407-bc0b454ca946-kube-api-access-jlgtv\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640916 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k8t2\" (UniqueName: \"kubernetes.io/projected/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc-kube-api-access-5k8t2\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.640929 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32514f48-c09d-4bfc-a407-bc0b454ca946-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.641521 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-logs" (OuterVolumeSpecName: "logs") pod "3c5f9b26-6c76-4af7-a811-d7d763f74ed6" (UID: "3c5f9b26-6c76-4af7-a811-d7d763f74ed6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.645632 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f707420-7a54-4f2c-ac75-38e09d8dfb61-logs" (OuterVolumeSpecName: "logs") pod "2f707420-7a54-4f2c-ac75-38e09d8dfb61" (UID: "2f707420-7a54-4f2c-ac75-38e09d8dfb61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.647565 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3c5f9b26-6c76-4af7-a811-d7d763f74ed6" (UID: "3c5f9b26-6c76-4af7-a811-d7d763f74ed6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.654900 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f707420-7a54-4f2c-ac75-38e09d8dfb61-kube-api-access-xn5tm" (OuterVolumeSpecName: "kube-api-access-xn5tm") pod "2f707420-7a54-4f2c-ac75-38e09d8dfb61" (UID: "2f707420-7a54-4f2c-ac75-38e09d8dfb61"). InnerVolumeSpecName "kube-api-access-xn5tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.661758 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-kube-api-access-52st8" (OuterVolumeSpecName: "kube-api-access-52st8") pod "3c5f9b26-6c76-4af7-a811-d7d763f74ed6" (UID: "3c5f9b26-6c76-4af7-a811-d7d763f74ed6"). InnerVolumeSpecName "kube-api-access-52st8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.669911 4733 generic.go:334] "Generic (PLEG): container finished" podID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerID="a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd" exitCode=0 Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.670011 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f757c77dd-bzp5x" event={"ID":"3c5f9b26-6c76-4af7-a811-d7d763f74ed6","Type":"ContainerDied","Data":"a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.670061 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f757c77dd-bzp5x" event={"ID":"3c5f9b26-6c76-4af7-a811-d7d763f74ed6","Type":"ContainerDied","Data":"d7f944db0d8b755d5a29260a295bbc66eb68e4af34fcc983a42e2ddc19c795cb"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.670086 4733 scope.go:117] "RemoveContainer" containerID="a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.670292 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f757c77dd-bzp5x" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.674474 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c8b0056-e347-4f24-aeea-20680c33bf06","Type":"ContainerDied","Data":"00657a27d84c4ba431e0def2a7ed219f0b276dd285ad90f3864cea5721e2fe7d"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.674582 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.685119 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron17fd-account-delete-pnhcm" event={"ID":"8b77d7a1-e0fe-466e-9629-ac6b5bab61dc","Type":"ContainerDied","Data":"909db43f49a047e2b4c76332e149c57629f691f23f6b6d720cf2e6be9ed4d49a"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.685167 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="909db43f49a047e2b4c76332e149c57629f691f23f6b6d720cf2e6be9ed4d49a" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.685226 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron17fd-account-delete-pnhcm" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.689463 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-config-data" (OuterVolumeSpecName: "config-data") pod "2f707420-7a54-4f2c-ac75-38e09d8dfb61" (UID: "2f707420-7a54-4f2c-ac75-38e09d8dfb61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.691574 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f707420-7a54-4f2c-ac75-38e09d8dfb61" (UID: "2f707420-7a54-4f2c-ac75-38e09d8dfb61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.691634 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance776f-account-delete-696d2" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.691654 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance776f-account-delete-696d2" event={"ID":"32514f48-c09d-4bfc-a407-bc0b454ca946","Type":"ContainerDied","Data":"c83599f693cb8a879141429622a802f59fe10e01e2c7c2563e439c9be03fc14a"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.691693 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c83599f693cb8a879141429622a802f59fe10e01e2c7c2563e439c9be03fc14a" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.698184 4733 scope.go:117] "RemoveContainer" containerID="1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.702546 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"98545dd5-cfa2-4c6c-97e9-6985a1d5a60d","Type":"ContainerDied","Data":"554083f376278140ce3ece24f4bafc5e456bf9cc4252f50cdae16d531cec3aeb"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.702629 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.709053 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.709818 4733 generic.go:334] "Generic (PLEG): container finished" podID="3d28f771-2fba-42d9-8fc0-1107d17db527" containerID="5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054" exitCode=0 Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.709898 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.710124 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3d28f771-2fba-42d9-8fc0-1107d17db527","Type":"ContainerDied","Data":"5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.710145 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3d28f771-2fba-42d9-8fc0-1107d17db527","Type":"ContainerDied","Data":"f9c96f88c3e7c60ff1d0226643539b80718d7ba96846aeb939ea4ed7c48fa513"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.712452 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"658fbbbc-cb7d-4795-be06-76a46b589943","Type":"ContainerDied","Data":"1fd5839cf44a213dade16102c343e15d5eebb9a683216736c05d552a656294f6"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.712548 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.714623 4733 generic.go:334] "Generic (PLEG): container finished" podID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerID="f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e" exitCode=0 Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.714680 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd72019f-8a36-43af-8d73-7c3e804c1baa","Type":"ContainerDied","Data":"f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.714697 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dd72019f-8a36-43af-8d73-7c3e804c1baa","Type":"ContainerDied","Data":"9a3d6a26497154ee9585e27e15e7ef56316cdacda43fca2b6de46e9d06f61e93"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.714733 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.718581 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.723917 4733 generic.go:334] "Generic (PLEG): container finished" podID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerID="135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39" exitCode=0 Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.723969 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.724012 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f707420-7a54-4f2c-ac75-38e09d8dfb61","Type":"ContainerDied","Data":"135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.724053 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f707420-7a54-4f2c-ac75-38e09d8dfb61","Type":"ContainerDied","Data":"4b1d5f86b1d49b0f44bcb86cbce3bdb23b5b724a2ae48075989c0629e879f98e"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.725907 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c5f9b26-6c76-4af7-a811-d7d763f74ed6" (UID: "3c5f9b26-6c76-4af7-a811-d7d763f74ed6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.726035 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data" (OuterVolumeSpecName: "config-data") pod "3c5f9b26-6c76-4af7-a811-d7d763f74ed6" (UID: "3c5f9b26-6c76-4af7-a811-d7d763f74ed6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.728864 4733 scope.go:117] "RemoveContainer" containerID="a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd" Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.730280 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd\": container with ID starting with a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd not found: ID does not exist" containerID="a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.730327 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd"} err="failed to get container status \"a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd\": rpc error: code = NotFound desc = could not find container \"a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd\": container with ID starting with a2c3ef901aa8e76cef8c1425dcf49b694ee1be11486f77586df11114a6efdafd not found: ID does not exist" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.730347 4733 scope.go:117] "RemoveContainer" containerID="1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d" Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.730710 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d\": container with ID starting with 1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d not found: ID does not exist" containerID="1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.730728 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d"} err="failed to get container status \"1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d\": rpc error: code = NotFound desc = could not find container \"1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d\": container with ID starting with 1a186c68a271219281d4e642cd1b10ba770e4c64f5c6d34e6dc4f90eb337cb3d not found: ID does not exist" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.730740 4733 scope.go:117] "RemoveContainer" containerID="cbcf529109c0c4af1bdf871c346e94fb3c2778251cd7b3b1d1b5864a31229cb3" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.731374 4733 generic.go:334] "Generic (PLEG): container finished" podID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerID="cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4" exitCode=0 Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.731395 4733 generic.go:334] "Generic (PLEG): container finished" podID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerID="43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864" exitCode=0 Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.731658 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerDied","Data":"cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.731684 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerDied","Data":"43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864"} Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.731735 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.732431 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2f707420-7a54-4f2c-ac75-38e09d8dfb61" (UID: "2f707420-7a54-4f2c-ac75-38e09d8dfb61"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.733220 4733 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell06ac2-account-delete-rs267" secret="" err="secret \"galera-openstack-dockercfg-c4nlc\" not found" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.734932 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder7119-account-delete-kzkjb" podUID="3f1c1320-5093-464a-a9a0-8a4cb668d70f" containerName="mariadb-account-delete" containerID="cri-o://1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414" gracePeriod=30 Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742145 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-memcached-tls-certs\") pod \"3d28f771-2fba-42d9-8fc0-1107d17db527\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742201 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-config-data\") pod \"3d28f771-2fba-42d9-8fc0-1107d17db527\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742336 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-kolla-config\") pod \"3d28f771-2fba-42d9-8fc0-1107d17db527\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742436 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-combined-ca-bundle\") pod \"3d28f771-2fba-42d9-8fc0-1107d17db527\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742479 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gd8p\" (UniqueName: \"kubernetes.io/projected/3d28f771-2fba-42d9-8fc0-1107d17db527-kube-api-access-9gd8p\") pod \"3d28f771-2fba-42d9-8fc0-1107d17db527\" (UID: \"3d28f771-2fba-42d9-8fc0-1107d17db527\") " Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742836 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742851 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742860 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742868 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742877 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn5tm\" (UniqueName: \"kubernetes.io/projected/2f707420-7a54-4f2c-ac75-38e09d8dfb61-kube-api-access-xn5tm\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742887 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742895 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742903 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742912 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52st8\" (UniqueName: \"kubernetes.io/projected/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-kube-api-access-52st8\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.742920 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f707420-7a54-4f2c-ac75-38e09d8dfb61-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.743486 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-config-data" (OuterVolumeSpecName: "config-data") pod "3d28f771-2fba-42d9-8fc0-1107d17db527" (UID: "3d28f771-2fba-42d9-8fc0-1107d17db527"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.745468 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "3d28f771-2fba-42d9-8fc0-1107d17db527" (UID: "3d28f771-2fba-42d9-8fc0-1107d17db527"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.745474 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.756727 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3c5f9b26-6c76-4af7-a811-d7d763f74ed6" (UID: "3c5f9b26-6c76-4af7-a811-d7d763f74ed6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.758512 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2f707420-7a54-4f2c-ac75-38e09d8dfb61" (UID: "2f707420-7a54-4f2c-ac75-38e09d8dfb61"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.765754 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.766737 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d28f771-2fba-42d9-8fc0-1107d17db527-kube-api-access-9gd8p" (OuterVolumeSpecName: "kube-api-access-9gd8p") pod "3d28f771-2fba-42d9-8fc0-1107d17db527" (UID: "3d28f771-2fba-42d9-8fc0-1107d17db527"). InnerVolumeSpecName "kube-api-access-9gd8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.781816 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance776f-account-delete-696d2"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.786740 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance776f-account-delete-696d2"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.787867 4733 scope.go:117] "RemoveContainer" containerID="63006d63457f2e071c6d109c86778556fbf386e34e8d0116bd25dc63bd13509f" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.794350 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron17fd-account-delete-pnhcm"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.794947 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3c5f9b26-6c76-4af7-a811-d7d763f74ed6" (UID: "3c5f9b26-6c76-4af7-a811-d7d763f74ed6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.798265 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d28f771-2fba-42d9-8fc0-1107d17db527" (UID: "3d28f771-2fba-42d9-8fc0-1107d17db527"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.807610 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron17fd-account-delete-pnhcm"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.819279 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "3d28f771-2fba-42d9-8fc0-1107d17db527" (UID: "3d28f771-2fba-42d9-8fc0-1107d17db527"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.841119 4733 scope.go:117] "RemoveContainer" containerID="6560f5a95ae3f9b652dd82bb257bd795655e85bb32c02059e6da7c32d0b36e95" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.845335 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.845376 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.845386 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gd8p\" (UniqueName: \"kubernetes.io/projected/3d28f771-2fba-42d9-8fc0-1107d17db527-kube-api-access-9gd8p\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.845398 4733 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d28f771-2fba-42d9-8fc0-1107d17db527-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.845410 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.845422 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f707420-7a54-4f2c-ac75-38e09d8dfb61-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.845432 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f9b26-6c76-4af7-a811-d7d763f74ed6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.845440 4733 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d28f771-2fba-42d9-8fc0-1107d17db527-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.851176 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.865271 4733 scope.go:117] "RemoveContainer" containerID="5fbdddb85e7e76ccd5911f7a283bd09d9eae7deb69d967d864ce48da4e7ac1be" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.867741 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.873583 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.878457 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.883461 4733 scope.go:117] "RemoveContainer" containerID="5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.885887 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.893048 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.910345 4733 scope.go:117] "RemoveContainer" containerID="5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054" Dec 04 18:03:46 crc kubenswrapper[4733]: E1204 18:03:46.911311 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054\": container with ID starting with 5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054 not found: ID does not exist" containerID="5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.911374 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054"} err="failed to get container status \"5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054\": rpc error: code = NotFound desc = could not find container \"5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054\": container with ID starting with 5e4ba3289e39017fdf54f0324cfcd0569910b0be0a01431ff00a740875aad054 not found: ID does not exist" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.911401 4733 scope.go:117] "RemoveContainer" containerID="bd700eb5d3983184b0c921d502c52f9cada1865ba48dea23580767a8dbb28917" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.954894 4733 scope.go:117] "RemoveContainer" containerID="f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e" Dec 04 18:03:46 crc kubenswrapper[4733]: I1204 18:03:46.987366 4733 scope.go:117] "RemoveContainer" containerID="39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.196189 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.205732 4733 scope.go:117] "RemoveContainer" containerID="f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e" Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.208276 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e\": container with ID starting with f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e not found: ID does not exist" containerID="f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.208326 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e"} err="failed to get container status \"f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e\": rpc error: code = NotFound desc = could not find container \"f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e\": container with ID starting with f1bce0290097b5daa3870df9c2dbe251033a2e3d457a6172a3a82163fec6554e not found: ID does not exist" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.208356 4733 scope.go:117] "RemoveContainer" containerID="39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb" Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.210075 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb\": container with ID starting with 39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb not found: ID does not exist" containerID="39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.210115 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb"} err="failed to get container status \"39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb\": rpc error: code = NotFound desc = could not find container \"39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb\": container with ID starting with 39481fe47714e0c399b24f1488e0430c6dfbf12e7e5b2fb172147a9a8a5ff4cb not found: ID does not exist" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.210136 4733 scope.go:117] "RemoveContainer" containerID="135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.210080 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.235628 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.244215 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.247849 4733 scope.go:117] "RemoveContainer" containerID="53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd" Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.251977 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.252028 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts podName:32b94636-cf95-42f9-8589-e38b550053af nodeName:}" failed. No retries permitted until 2025-12-04 18:03:49.252014131 +0000 UTC m=+1491.207375177 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts") pod "novacell06ac2-account-delete-rs267" (UID: "32b94636-cf95-42f9-8589-e38b550053af") : configmap "openstack-scripts" not found Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.252348 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7f757c77dd-bzp5x"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.260459 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7f757c77dd-bzp5x"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.266438 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.271265 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.272767 4733 scope.go:117] "RemoveContainer" containerID="135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39" Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.273201 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39\": container with ID starting with 135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39 not found: ID does not exist" containerID="135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.273243 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39"} err="failed to get container status \"135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39\": rpc error: code = NotFound desc = could not find container \"135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39\": container with ID starting with 135ac749a0b7c4b0b3fb9d0934c050a4e5d11190d58267fd3899693872434c39 not found: ID does not exist" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.273883 4733 scope.go:117] "RemoveContainer" containerID="53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd" Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.274275 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd\": container with ID starting with 53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd not found: ID does not exist" containerID="53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.274346 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd"} err="failed to get container status \"53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd\": rpc error: code = NotFound desc = could not find container \"53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd\": container with ID starting with 53680f5e4710f81601e282c6f8ffa5e3470b8abdf5356f64f055329bc70e87fd not found: ID does not exist" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.353433 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/faf216c0-9cda-4f76-a4df-8855352db991-operator-scripts\") pod \"faf216c0-9cda-4f76-a4df-8855352db991\" (UID: \"faf216c0-9cda-4f76-a4df-8855352db991\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.353510 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6rss\" (UniqueName: \"kubernetes.io/projected/faf216c0-9cda-4f76-a4df-8855352db991-kube-api-access-z6rss\") pod \"faf216c0-9cda-4f76-a4df-8855352db991\" (UID: \"faf216c0-9cda-4f76-a4df-8855352db991\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.353593 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stmgx\" (UniqueName: \"kubernetes.io/projected/31d98e38-ee17-4974-94e7-6495564940be-kube-api-access-stmgx\") pod \"31d98e38-ee17-4974-94e7-6495564940be\" (UID: \"31d98e38-ee17-4974-94e7-6495564940be\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.353631 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d98e38-ee17-4974-94e7-6495564940be-operator-scripts\") pod \"31d98e38-ee17-4974-94e7-6495564940be\" (UID: \"31d98e38-ee17-4974-94e7-6495564940be\") " Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.354140 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.354206 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts podName:3f1c1320-5093-464a-a9a0-8a4cb668d70f nodeName:}" failed. No retries permitted until 2025-12-04 18:03:49.354189334 +0000 UTC m=+1491.309550390 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts") pod "cinder7119-account-delete-kzkjb" (UID: "3f1c1320-5093-464a-a9a0-8a4cb668d70f") : configmap "openstack-scripts" not found Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.354237 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faf216c0-9cda-4f76-a4df-8855352db991-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "faf216c0-9cda-4f76-a4df-8855352db991" (UID: "faf216c0-9cda-4f76-a4df-8855352db991"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.354631 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d98e38-ee17-4974-94e7-6495564940be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31d98e38-ee17-4974-94e7-6495564940be" (UID: "31d98e38-ee17-4974-94e7-6495564940be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.360009 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faf216c0-9cda-4f76-a4df-8855352db991-kube-api-access-z6rss" (OuterVolumeSpecName: "kube-api-access-z6rss") pod "faf216c0-9cda-4f76-a4df-8855352db991" (UID: "faf216c0-9cda-4f76-a4df-8855352db991"). InnerVolumeSpecName "kube-api-access-z6rss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.360051 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d98e38-ee17-4974-94e7-6495564940be-kube-api-access-stmgx" (OuterVolumeSpecName: "kube-api-access-stmgx") pod "31d98e38-ee17-4974-94e7-6495564940be" (UID: "31d98e38-ee17-4974-94e7-6495564940be"). InnerVolumeSpecName "kube-api-access-stmgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.379207 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="165c8851cb406979af0eb269716a0167a055f027eebbecd5ddc98f53e1a7d325" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.386833 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="165c8851cb406979af0eb269716a0167a055f027eebbecd5ddc98f53e1a7d325" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.394556 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="165c8851cb406979af0eb269716a0167a055f027eebbecd5ddc98f53e1a7d325" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.394592 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="75c5521d-9c39-4fa7-88ad-11eb375c6d9a" containerName="nova-scheduler-scheduler" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.441982 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.455901 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d98e38-ee17-4974-94e7-6495564940be-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.455930 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/faf216c0-9cda-4f76-a4df-8855352db991-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.455941 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6rss\" (UniqueName: \"kubernetes.io/projected/faf216c0-9cda-4f76-a4df-8855352db991-kube-api-access-z6rss\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.455949 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stmgx\" (UniqueName: \"kubernetes.io/projected/31d98e38-ee17-4974-94e7-6495564940be-kube-api-access-stmgx\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557089 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-kolla-config\") pod \"b9cb46d6-3619-4035-8431-0b8552c5e690\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557214 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"b9cb46d6-3619-4035-8431-0b8552c5e690\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557271 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-combined-ca-bundle\") pod \"b9cb46d6-3619-4035-8431-0b8552c5e690\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557330 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-generated\") pod \"b9cb46d6-3619-4035-8431-0b8552c5e690\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557411 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-operator-scripts\") pod \"b9cb46d6-3619-4035-8431-0b8552c5e690\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557448 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hb7p\" (UniqueName: \"kubernetes.io/projected/b9cb46d6-3619-4035-8431-0b8552c5e690-kube-api-access-8hb7p\") pod \"b9cb46d6-3619-4035-8431-0b8552c5e690\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557474 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-default\") pod \"b9cb46d6-3619-4035-8431-0b8552c5e690\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557508 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-galera-tls-certs\") pod \"b9cb46d6-3619-4035-8431-0b8552c5e690\" (UID: \"b9cb46d6-3619-4035-8431-0b8552c5e690\") " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557588 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "b9cb46d6-3619-4035-8431-0b8552c5e690" (UID: "b9cb46d6-3619-4035-8431-0b8552c5e690"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557946 4733 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.557974 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "b9cb46d6-3619-4035-8431-0b8552c5e690" (UID: "b9cb46d6-3619-4035-8431-0b8552c5e690"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.559325 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "b9cb46d6-3619-4035-8431-0b8552c5e690" (UID: "b9cb46d6-3619-4035-8431-0b8552c5e690"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.559405 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b9cb46d6-3619-4035-8431-0b8552c5e690" (UID: "b9cb46d6-3619-4035-8431-0b8552c5e690"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.564147 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9cb46d6-3619-4035-8431-0b8552c5e690-kube-api-access-8hb7p" (OuterVolumeSpecName: "kube-api-access-8hb7p") pod "b9cb46d6-3619-4035-8431-0b8552c5e690" (UID: "b9cb46d6-3619-4035-8431-0b8552c5e690"). InnerVolumeSpecName "kube-api-access-8hb7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.565937 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "b9cb46d6-3619-4035-8431-0b8552c5e690" (UID: "b9cb46d6-3619-4035-8431-0b8552c5e690"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.592743 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9cb46d6-3619-4035-8431-0b8552c5e690" (UID: "b9cb46d6-3619-4035-8431-0b8552c5e690"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.599216 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "b9cb46d6-3619-4035-8431-0b8552c5e690" (UID: "b9cb46d6-3619-4035-8431-0b8552c5e690"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.659569 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.659609 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.659625 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.659637 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.659647 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hb7p\" (UniqueName: \"kubernetes.io/projected/b9cb46d6-3619-4035-8431-0b8552c5e690-kube-api-access-8hb7p\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.659658 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b9cb46d6-3619-4035-8431-0b8552c5e690-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.659743 4733 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9cb46d6-3619-4035-8431-0b8552c5e690-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.680913 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.744607 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi5c0f-account-delete-zk9rz" event={"ID":"faf216c0-9cda-4f76-a4df-8855352db991","Type":"ContainerDied","Data":"2a42bb3a9d25560a029b13a943a763cdfe7681f5633755f2a2552c368888e0c6"} Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.744668 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a42bb3a9d25560a029b13a943a763cdfe7681f5633755f2a2552c368888e0c6" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.744751 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi5c0f-account-delete-zk9rz" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.755192 4733 generic.go:334] "Generic (PLEG): container finished" podID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerID="e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c" exitCode=0 Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.755235 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9cb46d6-3619-4035-8431-0b8552c5e690","Type":"ContainerDied","Data":"e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c"} Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.755252 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9cb46d6-3619-4035-8431-0b8552c5e690","Type":"ContainerDied","Data":"c3601ba38f79f631e58366ddf941e47700a1f4d3e180b9a182e1776952c9f6a9"} Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.755268 4733 scope.go:117] "RemoveContainer" containerID="e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.755341 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.764024 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.772280 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican2aa9-account-delete-7bq8x" event={"ID":"31d98e38-ee17-4974-94e7-6495564940be","Type":"ContainerDied","Data":"6d53b5e0b7d789cf409282116070d1e8faedf4c50aabf9aaab81037888ca54c4"} Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.772313 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican2aa9-account-delete-7bq8x" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.772360 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell06ac2-account-delete-rs267" podUID="32b94636-cf95-42f9-8589-e38b550053af" containerName="mariadb-account-delete" containerID="cri-o://ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7" gracePeriod=30 Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.798868 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.807678 4733 scope.go:117] "RemoveContainer" containerID="f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.808897 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.820708 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi5c0f-account-delete-zk9rz"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.828608 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi5c0f-account-delete-zk9rz"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.834412 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican2aa9-account-delete-7bq8x"] Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.838464 4733 scope.go:117] "RemoveContainer" containerID="e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c" Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.838803 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c\": container with ID starting with e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c not found: ID does not exist" containerID="e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.838833 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c"} err="failed to get container status \"e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c\": rpc error: code = NotFound desc = could not find container \"e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c\": container with ID starting with e05fa0cc5577199d04c26564426da935dc3f5d7657b2e7e95adc97a9e88f3b3c not found: ID does not exist" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.838853 4733 scope.go:117] "RemoveContainer" containerID="f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.839315 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican2aa9-account-delete-7bq8x"] Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.839359 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6\": container with ID starting with f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6 not found: ID does not exist" containerID="f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.839377 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6"} err="failed to get container status \"f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6\": rpc error: code = NotFound desc = could not find container \"f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6\": container with ID starting with f717b94c561c393d59331fd110357f141a221af743d16c3b037001fccc4c99f6 not found: ID does not exist" Dec 04 18:03:47 crc kubenswrapper[4733]: I1204 18:03:47.839390 4733 scope.go:117] "RemoveContainer" containerID="c92d8d396adac2f52e75b63db8e67c9fd57abc43548e572275f0caf7dd9fb0ad" Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.967380 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:47 crc kubenswrapper[4733]: E1204 18:03:47.967468 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data podName:5e605c62-64b4-4417-80bb-bc3387881f7a nodeName:}" failed. No retries permitted until 2025-12-04 18:03:55.967452698 +0000 UTC m=+1497.922813744 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data") pod "rabbitmq-cell1-server-0" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a") : configmap "rabbitmq-cell1-config-data" not found Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.061881 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.063192 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.064443 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.064476 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="5eecf779-672e-4469-a134-9ecb0ef1bdde" containerName="nova-cell0-conductor-conductor" Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.192673 4733 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.192744 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data podName:1f55e362-d59c-4269-92c3-d5ca014a2ef1 nodeName:}" failed. No retries permitted until 2025-12-04 18:03:56.192728719 +0000 UTC m=+1498.148089765 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data") pod "rabbitmq-server-0" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1") : configmap "rabbitmq-config-data" not found Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.294001 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.294071 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts podName:08cdf00b-cf60-4e3a-9283-833e3a8d95bf nodeName:}" failed. No retries permitted until 2025-12-04 18:03:52.294056752 +0000 UTC m=+1494.249417798 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts") pod "placementee67-account-delete-sxlxx" (UID: "08cdf00b-cf60-4e3a-9283-833e3a8d95bf") : configmap "openstack-scripts" not found Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.346773 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" path="/var/lib/kubelet/pods/12b58db1-9574-4081-a3c1-2853f003fa8f/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.347678 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" path="/var/lib/kubelet/pods/2c8b0056-e347-4f24-aeea-20680c33bf06/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.348292 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" path="/var/lib/kubelet/pods/2f707420-7a54-4f2c-ac75-38e09d8dfb61/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.349504 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d98e38-ee17-4974-94e7-6495564940be" path="/var/lib/kubelet/pods/31d98e38-ee17-4974-94e7-6495564940be/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.349956 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32514f48-c09d-4bfc-a407-bc0b454ca946" path="/var/lib/kubelet/pods/32514f48-c09d-4bfc-a407-bc0b454ca946/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.350510 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" path="/var/lib/kubelet/pods/3c5f9b26-6c76-4af7-a811-d7d763f74ed6/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.351486 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d28f771-2fba-42d9-8fc0-1107d17db527" path="/var/lib/kubelet/pods/3d28f771-2fba-42d9-8fc0-1107d17db527/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.351941 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="658fbbbc-cb7d-4795-be06-76a46b589943" path="/var/lib/kubelet/pods/658fbbbc-cb7d-4795-be06-76a46b589943/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.352365 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b77d7a1-e0fe-466e-9629-ac6b5bab61dc" path="/var/lib/kubelet/pods/8b77d7a1-e0fe-466e-9629-ac6b5bab61dc/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.353444 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" path="/var/lib/kubelet/pods/98545dd5-cfa2-4c6c-97e9-6985a1d5a60d/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.354115 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" path="/var/lib/kubelet/pods/b9cb46d6-3619-4035-8431-0b8552c5e690/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.355269 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" path="/var/lib/kubelet/pods/dd72019f-8a36-43af-8d73-7c3e804c1baa/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.355929 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faf216c0-9cda-4f76-a4df-8855352db991" path="/var/lib/kubelet/pods/faf216c0-9cda-4f76-a4df-8855352db991/volumes" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.501088 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597625 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-server-conf\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597709 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-erlang-cookie\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597744 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h65g\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-kube-api-access-8h65g\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597763 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597785 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597842 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-plugins-conf\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597872 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-plugins\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597906 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-tls\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597942 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-confd\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597965 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e605c62-64b4-4417-80bb-bc3387881f7a-erlang-cookie-secret\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.597988 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e605c62-64b4-4417-80bb-bc3387881f7a-pod-info\") pod \"5e605c62-64b4-4417-80bb-bc3387881f7a\" (UID: \"5e605c62-64b4-4417-80bb-bc3387881f7a\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.598884 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.599404 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.603231 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5e605c62-64b4-4417-80bb-bc3387881f7a-pod-info" (OuterVolumeSpecName: "pod-info") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.603497 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.605642 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.606870 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e605c62-64b4-4417-80bb-bc3387881f7a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.606941 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-kube-api-access-8h65g" (OuterVolumeSpecName: "kube-api-access-8h65g") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "kube-api-access-8h65g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.608348 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.628498 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data" (OuterVolumeSpecName: "config-data") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.649543 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-server-conf" (OuterVolumeSpecName: "server-conf") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699295 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699330 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h65g\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-kube-api-access-8h65g\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699361 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699373 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699383 4733 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699393 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699402 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699411 4733 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e605c62-64b4-4417-80bb-bc3387881f7a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699422 4733 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e605c62-64b4-4417-80bb-bc3387881f7a-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.699431 4733 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e605c62-64b4-4417-80bb-bc3387881f7a-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.707931 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.728532 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5e605c62-64b4-4417-80bb-bc3387881f7a" (UID: "5e605c62-64b4-4417-80bb-bc3387881f7a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.772778 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800124 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f55e362-d59c-4269-92c3-d5ca014a2ef1-erlang-cookie-secret\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800173 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800206 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-plugins-conf\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800228 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-erlang-cookie\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800247 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800268 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm2ds\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-kube-api-access-hm2ds\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800287 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-plugins\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800321 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-server-conf\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800344 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f55e362-d59c-4269-92c3-d5ca014a2ef1-pod-info\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800363 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-confd\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800408 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-tls\") pod \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\" (UID: \"1f55e362-d59c-4269-92c3-d5ca014a2ef1\") " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800900 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.800916 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e605c62-64b4-4417-80bb-bc3387881f7a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.802222 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.802241 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.802670 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.805976 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-kube-api-access-hm2ds" (OuterVolumeSpecName: "kube-api-access-hm2ds") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "kube-api-access-hm2ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.805999 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.807424 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.811160 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f55e362-d59c-4269-92c3-d5ca014a2ef1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.816118 4733 generic.go:334] "Generic (PLEG): container finished" podID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerID="d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d" exitCode=0 Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.816186 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e605c62-64b4-4417-80bb-bc3387881f7a","Type":"ContainerDied","Data":"d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d"} Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.816213 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e605c62-64b4-4417-80bb-bc3387881f7a","Type":"ContainerDied","Data":"9081bacac85505f2a714990eec713b64d0c250ae5e824623e7cd9d39326a3676"} Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.816230 4733 scope.go:117] "RemoveContainer" containerID="d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.816316 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.817901 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1f55e362-d59c-4269-92c3-d5ca014a2ef1-pod-info" (OuterVolumeSpecName: "pod-info") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.821111 4733 generic.go:334] "Generic (PLEG): container finished" podID="9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" containerID="912469a5b6cd827fb51c283e6484beba8f59e32cbc52ed716d042adc91ed7c90" exitCode=0 Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.821158 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64bd4cdd64-fftgx" event={"ID":"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa","Type":"ContainerDied","Data":"912469a5b6cd827fb51c283e6484beba8f59e32cbc52ed716d042adc91ed7c90"} Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.823860 4733 generic.go:334] "Generic (PLEG): container finished" podID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerID="98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25" exitCode=0 Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.823909 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f55e362-d59c-4269-92c3-d5ca014a2ef1","Type":"ContainerDied","Data":"98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25"} Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.823936 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f55e362-d59c-4269-92c3-d5ca014a2ef1","Type":"ContainerDied","Data":"53a625c56c3f13329a1fe71e23d20de407a9ea740aa83d404bad031a89416a66"} Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.823999 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.835243 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data" (OuterVolumeSpecName: "config-data") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.842867 4733 scope.go:117] "RemoveContainer" containerID="9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.855059 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-server-conf" (OuterVolumeSpecName: "server-conf") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.870534 4733 scope.go:117] "RemoveContainer" containerID="d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d" Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.889692 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d\": container with ID starting with d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d not found: ID does not exist" containerID="d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.889743 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d"} err="failed to get container status \"d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d\": rpc error: code = NotFound desc = could not find container \"d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d\": container with ID starting with d69cfc35a99193a87fdf13fb719fe62769254ff7e5d4f3e94238391ed091538d not found: ID does not exist" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.889817 4733 scope.go:117] "RemoveContainer" containerID="9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab" Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.890224 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab\": container with ID starting with 9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab not found: ID does not exist" containerID="9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.890282 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab"} err="failed to get container status \"9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab\": rpc error: code = NotFound desc = could not find container \"9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab\": container with ID starting with 9a0345c69e4f94a37adee81049143e27aceda49d70ed6790c9ba9f22f0e7d8ab not found: ID does not exist" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.890317 4733 scope.go:117] "RemoveContainer" containerID="98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902629 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902664 4733 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f55e362-d59c-4269-92c3-d5ca014a2ef1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902675 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902682 4733 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902692 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902719 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902727 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm2ds\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-kube-api-access-hm2ds\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902736 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902744 4733 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f55e362-d59c-4269-92c3-d5ca014a2ef1-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.902751 4733 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f55e362-d59c-4269-92c3-d5ca014a2ef1-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.910825 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.912911 4733 scope.go:117] "RemoveContainer" containerID="d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.918468 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.920916 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.928036 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1f55e362-d59c-4269-92c3-d5ca014a2ef1" (UID: "1f55e362-d59c-4269-92c3-d5ca014a2ef1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.931924 4733 scope.go:117] "RemoveContainer" containerID="98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25" Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.932613 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25\": container with ID starting with 98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25 not found: ID does not exist" containerID="98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.932663 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25"} err="failed to get container status \"98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25\": rpc error: code = NotFound desc = could not find container \"98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25\": container with ID starting with 98a0b052e81890087a2416614b642d231cea78c88d1c508ba30a5062d8673b25 not found: ID does not exist" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.932705 4733 scope.go:117] "RemoveContainer" containerID="d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce" Dec 04 18:03:48 crc kubenswrapper[4733]: E1204 18:03:48.933214 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce\": container with ID starting with d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce not found: ID does not exist" containerID="d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.933256 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce"} err="failed to get container status \"d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce\": rpc error: code = NotFound desc = could not find container \"d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce\": container with ID starting with d90c1d997236c9d118345f1aca557d7d6619709359a1f032eb068c6954adb7ce not found: ID does not exist" Dec 04 18:03:48 crc kubenswrapper[4733]: I1204 18:03:48.982810 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.009301 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.009337 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f55e362-d59c-4269-92c3-d5ca014a2ef1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.109783 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-credential-keys\") pod \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.109842 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-scripts\") pod \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.109876 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-internal-tls-certs\") pod \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.109990 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-fernet-keys\") pod \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.110034 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-config-data\") pod \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.110071 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-public-tls-certs\") pod \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.110124 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-combined-ca-bundle\") pod \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.110161 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2jgd\" (UniqueName: \"kubernetes.io/projected/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-kube-api-access-w2jgd\") pod \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\" (UID: \"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.113313 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" (UID: "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.115215 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-scripts" (OuterVolumeSpecName: "scripts") pod "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" (UID: "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.115292 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-kube-api-access-w2jgd" (OuterVolumeSpecName: "kube-api-access-w2jgd") pod "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" (UID: "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa"). InnerVolumeSpecName "kube-api-access-w2jgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.117522 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" (UID: "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.140593 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" (UID: "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.144271 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-config-data" (OuterVolumeSpecName: "config-data") pod "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" (UID: "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.161994 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" (UID: "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.177290 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" (UID: "9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.205096 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.210668 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.210690 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.211692 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.211717 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2jgd\" (UniqueName: \"kubernetes.io/projected/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-kube-api-access-w2jgd\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.211731 4733 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.211743 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.211753 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.211772 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.211782 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.211817 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.212516 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.212807 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.212842 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.213637 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.213672 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.274288 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.280943 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.313922 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.314029 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts podName:32b94636-cf95-42f9-8589-e38b550053af nodeName:}" failed. No retries permitted until 2025-12-04 18:03:53.313998027 +0000 UTC m=+1495.269359073 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts") pod "novacell06ac2-account-delete-rs267" (UID: "32b94636-cf95-42f9-8589-e38b550053af") : configmap "openstack-scripts" not found Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.415587 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.415710 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts podName:3f1c1320-5093-464a-a9a0-8a4cb668d70f nodeName:}" failed. No retries permitted until 2025-12-04 18:03:53.4156864 +0000 UTC m=+1495.371047496 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts") pod "cinder7119-account-delete-kzkjb" (UID: "3f1c1320-5093-464a-a9a0-8a4cb668d70f") : configmap "openstack-scripts" not found Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.685041 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.820227 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4z5m\" (UniqueName: \"kubernetes.io/projected/839d7537-ff12-4ee6-b36c-b192f130a6e4-kube-api-access-q4z5m\") pod \"839d7537-ff12-4ee6-b36c-b192f130a6e4\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.820270 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-combined-ca-bundle\") pod \"839d7537-ff12-4ee6-b36c-b192f130a6e4\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.820369 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-config-data\") pod \"839d7537-ff12-4ee6-b36c-b192f130a6e4\" (UID: \"839d7537-ff12-4ee6-b36c-b192f130a6e4\") " Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.824397 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/839d7537-ff12-4ee6-b36c-b192f130a6e4-kube-api-access-q4z5m" (OuterVolumeSpecName: "kube-api-access-q4z5m") pod "839d7537-ff12-4ee6-b36c-b192f130a6e4" (UID: "839d7537-ff12-4ee6-b36c-b192f130a6e4"). InnerVolumeSpecName "kube-api-access-q4z5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.841864 4733 generic.go:334] "Generic (PLEG): container finished" podID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerID="52c25be5cf28de218adebb584f7e6fe433e56d68d1b191f31c36c67cb6d7c1a1" exitCode=0 Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.841916 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" event={"ID":"c1de9dcb-fad2-485d-b59b-14646f9ddfdd","Type":"ContainerDied","Data":"52c25be5cf28de218adebb584f7e6fe433e56d68d1b191f31c36c67cb6d7c1a1"} Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.842968 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "839d7537-ff12-4ee6-b36c-b192f130a6e4" (UID: "839d7537-ff12-4ee6-b36c-b192f130a6e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.846527 4733 generic.go:334] "Generic (PLEG): container finished" podID="839d7537-ff12-4ee6-b36c-b192f130a6e4" containerID="bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f" exitCode=0 Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.846620 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.846606 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"839d7537-ff12-4ee6-b36c-b192f130a6e4","Type":"ContainerDied","Data":"bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f"} Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.846663 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"839d7537-ff12-4ee6-b36c-b192f130a6e4","Type":"ContainerDied","Data":"8ad19b34b1b777b6642864ee77eac4ec85d9d00ac31f2123a546d6df01a8a97b"} Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.846684 4733 scope.go:117] "RemoveContainer" containerID="bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.851604 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64bd4cdd64-fftgx" event={"ID":"9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa","Type":"ContainerDied","Data":"d401f46e77cc3710d349a561b1377b9d2083e6630dd963949302bf71530d8dfb"} Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.851658 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64bd4cdd64-fftgx" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.864582 4733 generic.go:334] "Generic (PLEG): container finished" podID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerID="594133e5106de2f14d436d478b5b6fa76670e3bcd1bb710de01b002b98c8b79a" exitCode=0 Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.864661 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-676cf9f69-6mwxl" event={"ID":"87f72b89-cea6-4243-8ec6-e1264b4901c3","Type":"ContainerDied","Data":"594133e5106de2f14d436d478b5b6fa76670e3bcd1bb710de01b002b98c8b79a"} Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.864683 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-676cf9f69-6mwxl" event={"ID":"87f72b89-cea6-4243-8ec6-e1264b4901c3","Type":"ContainerDied","Data":"ff5801d26152910887f23748ab78ea20d11f11f76e866284f98bf5c2f9f5b66a"} Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.864693 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff5801d26152910887f23748ab78ea20d11f11f76e866284f98bf5c2f9f5b66a" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.866161 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.867441 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-config-data" (OuterVolumeSpecName: "config-data") pod "839d7537-ff12-4ee6-b36c-b192f130a6e4" (UID: "839d7537-ff12-4ee6-b36c-b192f130a6e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.867821 4733 generic.go:334] "Generic (PLEG): container finished" podID="75c5521d-9c39-4fa7-88ad-11eb375c6d9a" containerID="165c8851cb406979af0eb269716a0167a055f027eebbecd5ddc98f53e1a7d325" exitCode=0 Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.867846 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"75c5521d-9c39-4fa7-88ad-11eb375c6d9a","Type":"ContainerDied","Data":"165c8851cb406979af0eb269716a0167a055f027eebbecd5ddc98f53e1a7d325"} Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.885745 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-64bd4cdd64-fftgx"] Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.896884 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-64bd4cdd64-fftgx"] Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.901359 4733 scope.go:117] "RemoveContainer" containerID="bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f" Dec 04 18:03:49 crc kubenswrapper[4733]: E1204 18:03:49.901849 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f\": container with ID starting with bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f not found: ID does not exist" containerID="bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.901885 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f"} err="failed to get container status \"bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f\": rpc error: code = NotFound desc = could not find container \"bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f\": container with ID starting with bf62bd1b057e279e0f71df38f59e7fb508370a9a53647b925c2713bec1595b5f not found: ID does not exist" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.901913 4733 scope.go:117] "RemoveContainer" containerID="912469a5b6cd827fb51c283e6484beba8f59e32cbc52ed716d042adc91ed7c90" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.921749 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4z5m\" (UniqueName: \"kubernetes.io/projected/839d7537-ff12-4ee6-b36c-b192f130a6e4-kube-api-access-q4z5m\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.921780 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:49 crc kubenswrapper[4733]: I1204 18:03:49.921857 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/839d7537-ff12-4ee6-b36c-b192f130a6e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.022510 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87f72b89-cea6-4243-8ec6-e1264b4901c3-logs\") pod \"87f72b89-cea6-4243-8ec6-e1264b4901c3\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.022561 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data\") pod \"87f72b89-cea6-4243-8ec6-e1264b4901c3\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.022596 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data-custom\") pod \"87f72b89-cea6-4243-8ec6-e1264b4901c3\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.022644 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-combined-ca-bundle\") pod \"87f72b89-cea6-4243-8ec6-e1264b4901c3\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.022770 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z478s\" (UniqueName: \"kubernetes.io/projected/87f72b89-cea6-4243-8ec6-e1264b4901c3-kube-api-access-z478s\") pod \"87f72b89-cea6-4243-8ec6-e1264b4901c3\" (UID: \"87f72b89-cea6-4243-8ec6-e1264b4901c3\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.023787 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87f72b89-cea6-4243-8ec6-e1264b4901c3-logs" (OuterVolumeSpecName: "logs") pod "87f72b89-cea6-4243-8ec6-e1264b4901c3" (UID: "87f72b89-cea6-4243-8ec6-e1264b4901c3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.027270 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87f72b89-cea6-4243-8ec6-e1264b4901c3-kube-api-access-z478s" (OuterVolumeSpecName: "kube-api-access-z478s") pod "87f72b89-cea6-4243-8ec6-e1264b4901c3" (UID: "87f72b89-cea6-4243-8ec6-e1264b4901c3"). InnerVolumeSpecName "kube-api-access-z478s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.027368 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "87f72b89-cea6-4243-8ec6-e1264b4901c3" (UID: "87f72b89-cea6-4243-8ec6-e1264b4901c3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.051054 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87f72b89-cea6-4243-8ec6-e1264b4901c3" (UID: "87f72b89-cea6-4243-8ec6-e1264b4901c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.095209 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data" (OuterVolumeSpecName: "config-data") pod "87f72b89-cea6-4243-8ec6-e1264b4901c3" (UID: "87f72b89-cea6-4243-8ec6-e1264b4901c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.124172 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z478s\" (UniqueName: \"kubernetes.io/projected/87f72b89-cea6-4243-8ec6-e1264b4901c3-kube-api-access-z478s\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.124207 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87f72b89-cea6-4243-8ec6-e1264b4901c3-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.124216 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.124226 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.124234 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f72b89-cea6-4243-8ec6-e1264b4901c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.145789 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.196380 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.196963 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.209890 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.243620 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.326168 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6sq9\" (UniqueName: \"kubernetes.io/projected/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-kube-api-access-c6sq9\") pod \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.326219 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data-custom\") pod \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.326274 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data\") pod \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.326309 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-combined-ca-bundle\") pod \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.326347 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-logs\") pod \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.326368 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-config-data\") pod \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.326439 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnpnl\" (UniqueName: \"kubernetes.io/projected/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-kube-api-access-rnpnl\") pod \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\" (UID: \"c1de9dcb-fad2-485d-b59b-14646f9ddfdd\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.326477 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-combined-ca-bundle\") pod \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\" (UID: \"75c5521d-9c39-4fa7-88ad-11eb375c6d9a\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.327213 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-logs" (OuterVolumeSpecName: "logs") pod "c1de9dcb-fad2-485d-b59b-14646f9ddfdd" (UID: "c1de9dcb-fad2-485d-b59b-14646f9ddfdd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.329635 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-kube-api-access-rnpnl" (OuterVolumeSpecName: "kube-api-access-rnpnl") pod "c1de9dcb-fad2-485d-b59b-14646f9ddfdd" (UID: "c1de9dcb-fad2-485d-b59b-14646f9ddfdd"). InnerVolumeSpecName "kube-api-access-rnpnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.330251 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c1de9dcb-fad2-485d-b59b-14646f9ddfdd" (UID: "c1de9dcb-fad2-485d-b59b-14646f9ddfdd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.334283 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-kube-api-access-c6sq9" (OuterVolumeSpecName: "kube-api-access-c6sq9") pod "75c5521d-9c39-4fa7-88ad-11eb375c6d9a" (UID: "75c5521d-9c39-4fa7-88ad-11eb375c6d9a"). InnerVolumeSpecName "kube-api-access-c6sq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.344874 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" path="/var/lib/kubelet/pods/1f55e362-d59c-4269-92c3-d5ca014a2ef1/volumes" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.345616 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" path="/var/lib/kubelet/pods/5e605c62-64b4-4417-80bb-bc3387881f7a/volumes" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.346632 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="839d7537-ff12-4ee6-b36c-b192f130a6e4" path="/var/lib/kubelet/pods/839d7537-ff12-4ee6-b36c-b192f130a6e4/volumes" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.347254 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" path="/var/lib/kubelet/pods/9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa/volumes" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.347845 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75c5521d-9c39-4fa7-88ad-11eb375c6d9a" (UID: "75c5521d-9c39-4fa7-88ad-11eb375c6d9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.356971 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-config-data" (OuterVolumeSpecName: "config-data") pod "75c5521d-9c39-4fa7-88ad-11eb375c6d9a" (UID: "75c5521d-9c39-4fa7-88ad-11eb375c6d9a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.365501 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1de9dcb-fad2-485d-b59b-14646f9ddfdd" (UID: "c1de9dcb-fad2-485d-b59b-14646f9ddfdd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.393372 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data" (OuterVolumeSpecName: "config-data") pod "c1de9dcb-fad2-485d-b59b-14646f9ddfdd" (UID: "c1de9dcb-fad2-485d-b59b-14646f9ddfdd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.428774 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d642r\" (UniqueName: \"kubernetes.io/projected/5eecf779-672e-4469-a134-9ecb0ef1bdde-kube-api-access-d642r\") pod \"5eecf779-672e-4469-a134-9ecb0ef1bdde\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429226 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-combined-ca-bundle\") pod \"5eecf779-672e-4469-a134-9ecb0ef1bdde\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429290 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-config-data\") pod \"5eecf779-672e-4469-a134-9ecb0ef1bdde\" (UID: \"5eecf779-672e-4469-a134-9ecb0ef1bdde\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429576 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429592 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6sq9\" (UniqueName: \"kubernetes.io/projected/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-kube-api-access-c6sq9\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429602 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429611 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429620 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429628 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-logs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429637 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75c5521d-9c39-4fa7-88ad-11eb375c6d9a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.429646 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnpnl\" (UniqueName: \"kubernetes.io/projected/c1de9dcb-fad2-485d-b59b-14646f9ddfdd-kube-api-access-rnpnl\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.433956 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eecf779-672e-4469-a134-9ecb0ef1bdde-kube-api-access-d642r" (OuterVolumeSpecName: "kube-api-access-d642r") pod "5eecf779-672e-4469-a134-9ecb0ef1bdde" (UID: "5eecf779-672e-4469-a134-9ecb0ef1bdde"). InnerVolumeSpecName "kube-api-access-d642r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.447715 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-config-data" (OuterVolumeSpecName: "config-data") pod "5eecf779-672e-4469-a134-9ecb0ef1bdde" (UID: "5eecf779-672e-4469-a134-9ecb0ef1bdde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.458156 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5eecf779-672e-4469-a134-9ecb0ef1bdde" (UID: "5eecf779-672e-4469-a134-9ecb0ef1bdde"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.530993 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d642r\" (UniqueName: \"kubernetes.io/projected/5eecf779-672e-4469-a134-9ecb0ef1bdde-kube-api-access-d642r\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.531033 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.531046 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecf779-672e-4469-a134-9ecb0ef1bdde-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.831335 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.878888 4733 generic.go:334] "Generic (PLEG): container finished" podID="5eecf779-672e-4469-a134-9ecb0ef1bdde" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" exitCode=0 Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.878996 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.879019 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5eecf779-672e-4469-a134-9ecb0ef1bdde","Type":"ContainerDied","Data":"96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b"} Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.879067 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5eecf779-672e-4469-a134-9ecb0ef1bdde","Type":"ContainerDied","Data":"7b4e7fd78c569c81d6798502728ecf4b61b99db55f050463555b435ea0c256e5"} Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.879085 4733 scope.go:117] "RemoveContainer" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.885657 4733 generic.go:334] "Generic (PLEG): container finished" podID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerID="aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a" exitCode=0 Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.885723 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerDied","Data":"aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a"} Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.885753 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90549281-a4f7-416e-92dd-cea0e94c9af7","Type":"ContainerDied","Data":"595538b422dbb0c27c9c270478e9f80ea12d525054e7920843d616e22bc29c48"} Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.885852 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.888688 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"75c5521d-9c39-4fa7-88ad-11eb375c6d9a","Type":"ContainerDied","Data":"4eab83c6136e711b66288a20c7055eea07d20bd55f58d36074cdb6a0d40733cc"} Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.888858 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.899468 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-676cf9f69-6mwxl" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.899755 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.900288 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75746f8666-mw9nb" event={"ID":"c1de9dcb-fad2-485d-b59b-14646f9ddfdd","Type":"ContainerDied","Data":"d65a8d949cb2ae790623066d55bbd6b2c4db7377b065fcd19fe33a27fb04c78d"} Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.917785 4733 scope.go:117] "RemoveContainer" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" Dec 04 18:03:50 crc kubenswrapper[4733]: E1204 18:03:50.926689 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b\": container with ID starting with 96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b not found: ID does not exist" containerID="96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.926810 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b"} err="failed to get container status \"96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b\": rpc error: code = NotFound desc = could not find container \"96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b\": container with ID starting with 96dd0c4ca1f0063e2a896654865d7ada058ea682a0f890d5e9505a5aec13424b not found: ID does not exist" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.926836 4733 scope.go:117] "RemoveContainer" containerID="cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.936217 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-scripts\") pod \"90549281-a4f7-416e-92dd-cea0e94c9af7\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.936256 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkb4b\" (UniqueName: \"kubernetes.io/projected/90549281-a4f7-416e-92dd-cea0e94c9af7-kube-api-access-qkb4b\") pod \"90549281-a4f7-416e-92dd-cea0e94c9af7\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.936338 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-combined-ca-bundle\") pod \"90549281-a4f7-416e-92dd-cea0e94c9af7\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.936364 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-config-data\") pod \"90549281-a4f7-416e-92dd-cea0e94c9af7\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.936401 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-log-httpd\") pod \"90549281-a4f7-416e-92dd-cea0e94c9af7\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.936471 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-ceilometer-tls-certs\") pod \"90549281-a4f7-416e-92dd-cea0e94c9af7\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.936549 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-run-httpd\") pod \"90549281-a4f7-416e-92dd-cea0e94c9af7\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.936586 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-sg-core-conf-yaml\") pod \"90549281-a4f7-416e-92dd-cea0e94c9af7\" (UID: \"90549281-a4f7-416e-92dd-cea0e94c9af7\") " Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.939627 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "90549281-a4f7-416e-92dd-cea0e94c9af7" (UID: "90549281-a4f7-416e-92dd-cea0e94c9af7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.940496 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "90549281-a4f7-416e-92dd-cea0e94c9af7" (UID: "90549281-a4f7-416e-92dd-cea0e94c9af7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.940756 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.943166 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90549281-a4f7-416e-92dd-cea0e94c9af7-kube-api-access-qkb4b" (OuterVolumeSpecName: "kube-api-access-qkb4b") pod "90549281-a4f7-416e-92dd-cea0e94c9af7" (UID: "90549281-a4f7-416e-92dd-cea0e94c9af7"). InnerVolumeSpecName "kube-api-access-qkb4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.943269 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-scripts" (OuterVolumeSpecName: "scripts") pod "90549281-a4f7-416e-92dd-cea0e94c9af7" (UID: "90549281-a4f7-416e-92dd-cea0e94c9af7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.949174 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.954167 4733 scope.go:117] "RemoveContainer" containerID="8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.965439 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "90549281-a4f7-416e-92dd-cea0e94c9af7" (UID: "90549281-a4f7-416e-92dd-cea0e94c9af7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:50 crc kubenswrapper[4733]: I1204 18:03:50.975820 4733 scope.go:117] "RemoveContainer" containerID="aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.018614 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-676cf9f69-6mwxl"] Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.018808 4733 scope.go:117] "RemoveContainer" containerID="43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.021037 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-676cf9f69-6mwxl"] Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.024587 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "90549281-a4f7-416e-92dd-cea0e94c9af7" (UID: "90549281-a4f7-416e-92dd-cea0e94c9af7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.030913 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-75746f8666-mw9nb"] Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.059622 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.059666 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.059678 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.059687 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkb4b\" (UniqueName: \"kubernetes.io/projected/90549281-a4f7-416e-92dd-cea0e94c9af7-kube-api-access-qkb4b\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.059701 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90549281-a4f7-416e-92dd-cea0e94c9af7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.059709 4733 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.059734 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-75746f8666-mw9nb"] Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.061537 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90549281-a4f7-416e-92dd-cea0e94c9af7" (UID: "90549281-a4f7-416e-92dd-cea0e94c9af7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.085670 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.092879 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-config-data" (OuterVolumeSpecName: "config-data") pod "90549281-a4f7-416e-92dd-cea0e94c9af7" (UID: "90549281-a4f7-416e-92dd-cea0e94c9af7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.098892 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.130115 4733 scope.go:117] "RemoveContainer" containerID="cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4" Dec 04 18:03:51 crc kubenswrapper[4733]: E1204 18:03:51.130847 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4\": container with ID starting with cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4 not found: ID does not exist" containerID="cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.130896 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4"} err="failed to get container status \"cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4\": rpc error: code = NotFound desc = could not find container \"cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4\": container with ID starting with cf96dfa07a8834f596e54cabc67f1dd234097798dba5b64a3af86a6aee93e3f4 not found: ID does not exist" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.130929 4733 scope.go:117] "RemoveContainer" containerID="8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7" Dec 04 18:03:51 crc kubenswrapper[4733]: E1204 18:03:51.131568 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7\": container with ID starting with 8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7 not found: ID does not exist" containerID="8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.131683 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7"} err="failed to get container status \"8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7\": rpc error: code = NotFound desc = could not find container \"8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7\": container with ID starting with 8867311ed6bcca89fe01f514728f017df9612ca80499bd47693a2ea419b34ab7 not found: ID does not exist" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.131758 4733 scope.go:117] "RemoveContainer" containerID="aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a" Dec 04 18:03:51 crc kubenswrapper[4733]: E1204 18:03:51.132301 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a\": container with ID starting with aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a not found: ID does not exist" containerID="aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.132362 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a"} err="failed to get container status \"aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a\": rpc error: code = NotFound desc = could not find container \"aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a\": container with ID starting with aaa82c7c2c86da85638cd5c26db0663ed386bcfca1d237253382cecacc83fb8a not found: ID does not exist" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.132379 4733 scope.go:117] "RemoveContainer" containerID="43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864" Dec 04 18:03:51 crc kubenswrapper[4733]: E1204 18:03:51.132616 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864\": container with ID starting with 43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864 not found: ID does not exist" containerID="43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.132644 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864"} err="failed to get container status \"43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864\": rpc error: code = NotFound desc = could not find container \"43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864\": container with ID starting with 43261b1691d71b5550a7889f7d23aa038bb7f2d6fdd7e8dede5be1a65962c864 not found: ID does not exist" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.132658 4733 scope.go:117] "RemoveContainer" containerID="165c8851cb406979af0eb269716a0167a055f027eebbecd5ddc98f53e1a7d325" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.149937 4733 scope.go:117] "RemoveContainer" containerID="52c25be5cf28de218adebb584f7e6fe433e56d68d1b191f31c36c67cb6d7c1a1" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.161682 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.161714 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90549281-a4f7-416e-92dd-cea0e94c9af7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.174549 4733 scope.go:117] "RemoveContainer" containerID="b2b953dbdaed812d6f01b224d3af9906a0d0da35f8c6eaeed44b254649426640" Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.218485 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:03:51 crc kubenswrapper[4733]: I1204 18:03:51.224153 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 18:03:52 crc kubenswrapper[4733]: E1204 18:03:52.297918 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:52 crc kubenswrapper[4733]: E1204 18:03:52.298035 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts podName:08cdf00b-cf60-4e3a-9283-833e3a8d95bf nodeName:}" failed. No retries permitted until 2025-12-04 18:04:00.298007402 +0000 UTC m=+1502.253368478 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts") pod "placementee67-account-delete-sxlxx" (UID: "08cdf00b-cf60-4e3a-9283-833e3a8d95bf") : configmap "openstack-scripts" not found Dec 04 18:03:52 crc kubenswrapper[4733]: I1204 18:03:52.345688 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eecf779-672e-4469-a134-9ecb0ef1bdde" path="/var/lib/kubelet/pods/5eecf779-672e-4469-a134-9ecb0ef1bdde/volumes" Dec 04 18:03:52 crc kubenswrapper[4733]: I1204 18:03:52.346226 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75c5521d-9c39-4fa7-88ad-11eb375c6d9a" path="/var/lib/kubelet/pods/75c5521d-9c39-4fa7-88ad-11eb375c6d9a/volumes" Dec 04 18:03:52 crc kubenswrapper[4733]: I1204 18:03:52.346751 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" path="/var/lib/kubelet/pods/87f72b89-cea6-4243-8ec6-e1264b4901c3/volumes" Dec 04 18:03:52 crc kubenswrapper[4733]: I1204 18:03:52.347820 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" path="/var/lib/kubelet/pods/90549281-a4f7-416e-92dd-cea0e94c9af7/volumes" Dec 04 18:03:52 crc kubenswrapper[4733]: I1204 18:03:52.348575 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" path="/var/lib/kubelet/pods/c1de9dcb-fad2-485d-b59b-14646f9ddfdd/volumes" Dec 04 18:03:53 crc kubenswrapper[4733]: E1204 18:03:53.414995 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:53 crc kubenswrapper[4733]: E1204 18:03:53.415067 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts podName:32b94636-cf95-42f9-8589-e38b550053af nodeName:}" failed. No retries permitted until 2025-12-04 18:04:01.415051168 +0000 UTC m=+1503.370412214 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts") pod "novacell06ac2-account-delete-rs267" (UID: "32b94636-cf95-42f9-8589-e38b550053af") : configmap "openstack-scripts" not found Dec 04 18:03:53 crc kubenswrapper[4733]: E1204 18:03:53.516634 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:03:53 crc kubenswrapper[4733]: E1204 18:03:53.516716 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts podName:3f1c1320-5093-464a-a9a0-8a4cb668d70f nodeName:}" failed. No retries permitted until 2025-12-04 18:04:01.516696278 +0000 UTC m=+1503.472057324 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts") pod "cinder7119-account-delete-kzkjb" (UID: "3f1c1320-5093-464a-a9a0-8a4cb668d70f") : configmap "openstack-scripts" not found Dec 04 18:03:54 crc kubenswrapper[4733]: E1204 18:03:54.218027 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:54 crc kubenswrapper[4733]: E1204 18:03:54.220279 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:54 crc kubenswrapper[4733]: E1204 18:03:54.221847 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:54 crc kubenswrapper[4733]: E1204 18:03:54.221928 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:03:54 crc kubenswrapper[4733]: E1204 18:03:54.233736 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:54 crc kubenswrapper[4733]: E1204 18:03:54.241347 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:54 crc kubenswrapper[4733]: E1204 18:03:54.246616 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:54 crc kubenswrapper[4733]: E1204 18:03:54.246731 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:03:57 crc kubenswrapper[4733]: I1204 18:03:57.998424 4733 generic.go:334] "Generic (PLEG): container finished" podID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerID="5a96bd3fd257dc89a4415f7d6d660060c72c788e7620add6273636ed06ed77b1" exitCode=0 Dec 04 18:03:57 crc kubenswrapper[4733]: I1204 18:03:57.998546 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84697bbbd9-9fltw" event={"ID":"4a9d9f43-f063-47fc-83b9-213c6ab06ef4","Type":"ContainerDied","Data":"5a96bd3fd257dc89a4415f7d6d660060c72c788e7620add6273636ed06ed77b1"} Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.431154 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.523181 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-public-tls-certs\") pod \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.523264 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-combined-ca-bundle\") pod \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.523293 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqsz6\" (UniqueName: \"kubernetes.io/projected/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-kube-api-access-hqsz6\") pod \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.523317 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-internal-tls-certs\") pod \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.523344 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-httpd-config\") pod \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.523383 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-ovndb-tls-certs\") pod \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.523403 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-config\") pod \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\" (UID: \"4a9d9f43-f063-47fc-83b9-213c6ab06ef4\") " Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.538667 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4a9d9f43-f063-47fc-83b9-213c6ab06ef4" (UID: "4a9d9f43-f063-47fc-83b9-213c6ab06ef4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.538752 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-kube-api-access-hqsz6" (OuterVolumeSpecName: "kube-api-access-hqsz6") pod "4a9d9f43-f063-47fc-83b9-213c6ab06ef4" (UID: "4a9d9f43-f063-47fc-83b9-213c6ab06ef4"). InnerVolumeSpecName "kube-api-access-hqsz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.572635 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4a9d9f43-f063-47fc-83b9-213c6ab06ef4" (UID: "4a9d9f43-f063-47fc-83b9-213c6ab06ef4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.591953 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a9d9f43-f063-47fc-83b9-213c6ab06ef4" (UID: "4a9d9f43-f063-47fc-83b9-213c6ab06ef4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.593003 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4a9d9f43-f063-47fc-83b9-213c6ab06ef4" (UID: "4a9d9f43-f063-47fc-83b9-213c6ab06ef4"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.598886 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-config" (OuterVolumeSpecName: "config") pod "4a9d9f43-f063-47fc-83b9-213c6ab06ef4" (UID: "4a9d9f43-f063-47fc-83b9-213c6ab06ef4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.599086 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4a9d9f43-f063-47fc-83b9-213c6ab06ef4" (UID: "4a9d9f43-f063-47fc-83b9-213c6ab06ef4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.625519 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.625646 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqsz6\" (UniqueName: \"kubernetes.io/projected/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-kube-api-access-hqsz6\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.625707 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.625759 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.625841 4733 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.625918 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-config\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:58 crc kubenswrapper[4733]: I1204 18:03:58.625977 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d9f43-f063-47fc-83b9-213c6ab06ef4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 18:03:59 crc kubenswrapper[4733]: I1204 18:03:59.014114 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84697bbbd9-9fltw" event={"ID":"4a9d9f43-f063-47fc-83b9-213c6ab06ef4","Type":"ContainerDied","Data":"9bca0ad92f9667c2fd0009160b8cb4c642f40b2ed8fbf5b7ef8a90fb36e088b5"} Dec 04 18:03:59 crc kubenswrapper[4733]: I1204 18:03:59.014200 4733 scope.go:117] "RemoveContainer" containerID="04763b048c7378462084dba4301a110daf2dd03bc1d5722e8ceee83b2170fcc9" Dec 04 18:03:59 crc kubenswrapper[4733]: I1204 18:03:59.014147 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84697bbbd9-9fltw" Dec 04 18:03:59 crc kubenswrapper[4733]: I1204 18:03:59.057243 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84697bbbd9-9fltw"] Dec 04 18:03:59 crc kubenswrapper[4733]: I1204 18:03:59.064658 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-84697bbbd9-9fltw"] Dec 04 18:03:59 crc kubenswrapper[4733]: I1204 18:03:59.068249 4733 scope.go:117] "RemoveContainer" containerID="5a96bd3fd257dc89a4415f7d6d660060c72c788e7620add6273636ed06ed77b1" Dec 04 18:03:59 crc kubenswrapper[4733]: E1204 18:03:59.203698 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:59 crc kubenswrapper[4733]: E1204 18:03:59.204732 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:59 crc kubenswrapper[4733]: E1204 18:03:59.205037 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:59 crc kubenswrapper[4733]: E1204 18:03:59.205179 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:03:59 crc kubenswrapper[4733]: E1204 18:03:59.205219 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:03:59 crc kubenswrapper[4733]: E1204 18:03:59.207206 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:59 crc kubenswrapper[4733]: E1204 18:03:59.209262 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:03:59 crc kubenswrapper[4733]: E1204 18:03:59.209327 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:04:00 crc kubenswrapper[4733]: I1204 18:04:00.349469 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" path="/var/lib/kubelet/pods/4a9d9f43-f063-47fc-83b9-213c6ab06ef4/volumes" Dec 04 18:04:00 crc kubenswrapper[4733]: E1204 18:04:00.355939 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:04:00 crc kubenswrapper[4733]: E1204 18:04:00.356033 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts podName:08cdf00b-cf60-4e3a-9283-833e3a8d95bf nodeName:}" failed. No retries permitted until 2025-12-04 18:04:16.356010375 +0000 UTC m=+1518.311371431 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts") pod "placementee67-account-delete-sxlxx" (UID: "08cdf00b-cf60-4e3a-9283-833e3a8d95bf") : configmap "openstack-scripts" not found Dec 04 18:04:01 crc kubenswrapper[4733]: E1204 18:04:01.505030 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:04:01 crc kubenswrapper[4733]: E1204 18:04:01.505124 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts podName:32b94636-cf95-42f9-8589-e38b550053af nodeName:}" failed. No retries permitted until 2025-12-04 18:04:17.505104898 +0000 UTC m=+1519.460465954 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts") pod "novacell06ac2-account-delete-rs267" (UID: "32b94636-cf95-42f9-8589-e38b550053af") : configmap "openstack-scripts" not found Dec 04 18:04:01 crc kubenswrapper[4733]: E1204 18:04:01.606846 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:04:01 crc kubenswrapper[4733]: E1204 18:04:01.606982 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts podName:3f1c1320-5093-464a-a9a0-8a4cb668d70f nodeName:}" failed. No retries permitted until 2025-12-04 18:04:17.606951954 +0000 UTC m=+1519.562313040 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts") pod "cinder7119-account-delete-kzkjb" (UID: "3f1c1320-5093-464a-a9a0-8a4cb668d70f") : configmap "openstack-scripts" not found Dec 04 18:04:04 crc kubenswrapper[4733]: E1204 18:04:04.203721 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:04:04 crc kubenswrapper[4733]: E1204 18:04:04.204836 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:04:04 crc kubenswrapper[4733]: E1204 18:04:04.205484 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:04:04 crc kubenswrapper[4733]: E1204 18:04:04.205552 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:04:04 crc kubenswrapper[4733]: E1204 18:04:04.205659 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:04:04 crc kubenswrapper[4733]: E1204 18:04:04.209330 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:04:04 crc kubenswrapper[4733]: E1204 18:04:04.211380 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:04:04 crc kubenswrapper[4733]: E1204 18:04:04.211455 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:04:09 crc kubenswrapper[4733]: E1204 18:04:09.204268 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:04:09 crc kubenswrapper[4733]: E1204 18:04:09.206056 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:04:09 crc kubenswrapper[4733]: E1204 18:04:09.206090 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:04:09 crc kubenswrapper[4733]: E1204 18:04:09.206970 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 18:04:09 crc kubenswrapper[4733]: E1204 18:04:09.207047 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:04:09 crc kubenswrapper[4733]: E1204 18:04:09.207483 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:04:09 crc kubenswrapper[4733]: E1204 18:04:09.208994 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 18:04:09 crc kubenswrapper[4733]: E1204 18:04:09.209064 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-rkmmv" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:04:10 crc kubenswrapper[4733]: I1204 18:04:10.967740 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rkmmv_0cf011d5-1c33-4ec6-8b1f-d1216a46d066/ovs-vswitchd/0.log" Dec 04 18:04:10 crc kubenswrapper[4733]: I1204 18:04:10.968940 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.073697 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-lib\") pod \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.073827 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-log\") pod \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.073870 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rprg6\" (UniqueName: \"kubernetes.io/projected/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-kube-api-access-rprg6\") pod \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.073884 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-etc-ovs\") pod \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.073913 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-run\") pod \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.073955 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-scripts\") pod \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\" (UID: \"0cf011d5-1c33-4ec6-8b1f-d1216a46d066\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.073984 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-lib" (OuterVolumeSpecName: "var-lib") pod "0cf011d5-1c33-4ec6-8b1f-d1216a46d066" (UID: "0cf011d5-1c33-4ec6-8b1f-d1216a46d066"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.074098 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-log" (OuterVolumeSpecName: "var-log") pod "0cf011d5-1c33-4ec6-8b1f-d1216a46d066" (UID: "0cf011d5-1c33-4ec6-8b1f-d1216a46d066"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.074148 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-run" (OuterVolumeSpecName: "var-run") pod "0cf011d5-1c33-4ec6-8b1f-d1216a46d066" (UID: "0cf011d5-1c33-4ec6-8b1f-d1216a46d066"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.074163 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "0cf011d5-1c33-4ec6-8b1f-d1216a46d066" (UID: "0cf011d5-1c33-4ec6-8b1f-d1216a46d066"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.074561 4733 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-lib\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.074594 4733 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.074614 4733 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.074631 4733 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.075436 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-scripts" (OuterVolumeSpecName: "scripts") pod "0cf011d5-1c33-4ec6-8b1f-d1216a46d066" (UID: "0cf011d5-1c33-4ec6-8b1f-d1216a46d066"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.089262 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-kube-api-access-rprg6" (OuterVolumeSpecName: "kube-api-access-rprg6") pod "0cf011d5-1c33-4ec6-8b1f-d1216a46d066" (UID: "0cf011d5-1c33-4ec6-8b1f-d1216a46d066"). InnerVolumeSpecName "kube-api-access-rprg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.153138 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rkmmv_0cf011d5-1c33-4ec6-8b1f-d1216a46d066/ovs-vswitchd/0.log" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.154203 4733 generic.go:334] "Generic (PLEG): container finished" podID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" exitCode=137 Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.154315 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rkmmv" event={"ID":"0cf011d5-1c33-4ec6-8b1f-d1216a46d066","Type":"ContainerDied","Data":"8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1"} Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.154355 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rkmmv" event={"ID":"0cf011d5-1c33-4ec6-8b1f-d1216a46d066","Type":"ContainerDied","Data":"ad730f55d153bbc15f9343dc147daeed6c5833d75283b532b5091379e2f35d53"} Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.154375 4733 scope.go:117] "RemoveContainer" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.154328 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rkmmv" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.175473 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rprg6\" (UniqueName: \"kubernetes.io/projected/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-kube-api-access-rprg6\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.175504 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0cf011d5-1c33-4ec6-8b1f-d1216a46d066-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.189512 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-rkmmv"] Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.191255 4733 scope.go:117] "RemoveContainer" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.195895 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-rkmmv"] Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.220008 4733 scope.go:117] "RemoveContainer" containerID="18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.252475 4733 scope.go:117] "RemoveContainer" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" Dec 04 18:04:11 crc kubenswrapper[4733]: E1204 18:04:11.253110 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1\": container with ID starting with 8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1 not found: ID does not exist" containerID="8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.253194 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1"} err="failed to get container status \"8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1\": rpc error: code = NotFound desc = could not find container \"8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1\": container with ID starting with 8130d4a261fbfea00fe54308d47ab2d5776c3ab9a9c91a3940e12424f2fa76a1 not found: ID does not exist" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.253287 4733 scope.go:117] "RemoveContainer" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" Dec 04 18:04:11 crc kubenswrapper[4733]: E1204 18:04:11.253919 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16\": container with ID starting with 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 not found: ID does not exist" containerID="102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.254006 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16"} err="failed to get container status \"102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16\": rpc error: code = NotFound desc = could not find container \"102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16\": container with ID starting with 102871d25d1ff00a411e8c768f6adbf14535b84bb6476d0e7d417f49523ebe16 not found: ID does not exist" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.254042 4733 scope.go:117] "RemoveContainer" containerID="18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc" Dec 04 18:04:11 crc kubenswrapper[4733]: E1204 18:04:11.255439 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc\": container with ID starting with 18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc not found: ID does not exist" containerID="18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.255490 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc"} err="failed to get container status \"18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc\": rpc error: code = NotFound desc = could not find container \"18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc\": container with ID starting with 18f3a5808db013d07bc3a6ff813d63b00da899d5d59b0d87061aa126422bcdbc not found: ID does not exist" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.869081 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.990776 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") pod \"75da96fa-8079-442d-8993-de5658a23f91\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.992120 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"75da96fa-8079-442d-8993-de5658a23f91\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.992242 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-lock\") pod \"75da96fa-8079-442d-8993-de5658a23f91\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.992299 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-cache\") pod \"75da96fa-8079-442d-8993-de5658a23f91\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.992396 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpxbm\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-kube-api-access-qpxbm\") pod \"75da96fa-8079-442d-8993-de5658a23f91\" (UID: \"75da96fa-8079-442d-8993-de5658a23f91\") " Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.993738 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-lock" (OuterVolumeSpecName: "lock") pod "75da96fa-8079-442d-8993-de5658a23f91" (UID: "75da96fa-8079-442d-8993-de5658a23f91"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.993872 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-cache" (OuterVolumeSpecName: "cache") pod "75da96fa-8079-442d-8993-de5658a23f91" (UID: "75da96fa-8079-442d-8993-de5658a23f91"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:04:11 crc kubenswrapper[4733]: I1204 18:04:11.998099 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "75da96fa-8079-442d-8993-de5658a23f91" (UID: "75da96fa-8079-442d-8993-de5658a23f91"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.000094 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-kube-api-access-qpxbm" (OuterVolumeSpecName: "kube-api-access-qpxbm") pod "75da96fa-8079-442d-8993-de5658a23f91" (UID: "75da96fa-8079-442d-8993-de5658a23f91"). InnerVolumeSpecName "kube-api-access-qpxbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.001535 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "swift") pod "75da96fa-8079-442d-8993-de5658a23f91" (UID: "75da96fa-8079-442d-8993-de5658a23f91"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.094713 4733 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.094780 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.094811 4733 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-lock\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.094823 4733 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/75da96fa-8079-442d-8993-de5658a23f91-cache\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.094835 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpxbm\" (UniqueName: \"kubernetes.io/projected/75da96fa-8079-442d-8993-de5658a23f91-kube-api-access-qpxbm\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.114495 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.175087 4733 generic.go:334] "Generic (PLEG): container finished" podID="75da96fa-8079-442d-8993-de5658a23f91" containerID="5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c" exitCode=137 Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.175163 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c"} Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.175191 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"75da96fa-8079-442d-8993-de5658a23f91","Type":"ContainerDied","Data":"fd17271bff52c7522d9ce0902111b6a3f9ff36c957ced644ec80c5590f8cacb9"} Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.175207 4733 scope.go:117] "RemoveContainer" containerID="5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.175262 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.196878 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.208214 4733 scope.go:117] "RemoveContainer" containerID="92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.224490 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.231816 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.244770 4733 scope.go:117] "RemoveContainer" containerID="b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.264633 4733 scope.go:117] "RemoveContainer" containerID="e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.283317 4733 scope.go:117] "RemoveContainer" containerID="33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.305012 4733 scope.go:117] "RemoveContainer" containerID="1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.330620 4733 scope.go:117] "RemoveContainer" containerID="afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.345587 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" path="/var/lib/kubelet/pods/0cf011d5-1c33-4ec6-8b1f-d1216a46d066/volumes" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.347260 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75da96fa-8079-442d-8993-de5658a23f91" path="/var/lib/kubelet/pods/75da96fa-8079-442d-8993-de5658a23f91/volumes" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.352855 4733 scope.go:117] "RemoveContainer" containerID="aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.374547 4733 scope.go:117] "RemoveContainer" containerID="a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.395486 4733 scope.go:117] "RemoveContainer" containerID="e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.413880 4733 scope.go:117] "RemoveContainer" containerID="e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.434372 4733 scope.go:117] "RemoveContainer" containerID="13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.458011 4733 scope.go:117] "RemoveContainer" containerID="cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.492965 4733 scope.go:117] "RemoveContainer" containerID="da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.529022 4733 scope.go:117] "RemoveContainer" containerID="1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.545401 4733 scope.go:117] "RemoveContainer" containerID="5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.545921 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c\": container with ID starting with 5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c not found: ID does not exist" containerID="5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.545957 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c"} err="failed to get container status \"5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c\": rpc error: code = NotFound desc = could not find container \"5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c\": container with ID starting with 5916a08d0999031d0412b39fa2194fd4ae1a28df0eabfe43ec9e391b7f6af17c not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.545982 4733 scope.go:117] "RemoveContainer" containerID="92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.546227 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c\": container with ID starting with 92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c not found: ID does not exist" containerID="92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.546259 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c"} err="failed to get container status \"92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c\": rpc error: code = NotFound desc = could not find container \"92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c\": container with ID starting with 92c5e025736c3441fca045734e93d957bead748e2771f48596e4faa0f8f4857c not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.546278 4733 scope.go:117] "RemoveContainer" containerID="b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.546838 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0\": container with ID starting with b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0 not found: ID does not exist" containerID="b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.546873 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0"} err="failed to get container status \"b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0\": rpc error: code = NotFound desc = could not find container \"b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0\": container with ID starting with b73453ba5b0ac08d54e32ad60fd83cd742c3fc4ff727d3e8a9725c276fce91a0 not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.546892 4733 scope.go:117] "RemoveContainer" containerID="e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.549112 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828\": container with ID starting with e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828 not found: ID does not exist" containerID="e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.549135 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828"} err="failed to get container status \"e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828\": rpc error: code = NotFound desc = could not find container \"e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828\": container with ID starting with e5413c05d69fd9c775162a1fb526dc2f46b187ba2572a8cc4f98557178abe828 not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.549150 4733 scope.go:117] "RemoveContainer" containerID="33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.549515 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f\": container with ID starting with 33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f not found: ID does not exist" containerID="33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.549578 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f"} err="failed to get container status \"33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f\": rpc error: code = NotFound desc = could not find container \"33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f\": container with ID starting with 33a89b13cbda14add83ae8bb52d02dc68ab39a06003a1deaba6ffc848eb33b6f not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.549619 4733 scope.go:117] "RemoveContainer" containerID="1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.549964 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda\": container with ID starting with 1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda not found: ID does not exist" containerID="1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.549986 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda"} err="failed to get container status \"1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda\": rpc error: code = NotFound desc = could not find container \"1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda\": container with ID starting with 1070e333a111f5f268eca2d64aa971cc3815fa08b484e2422e175e47d60e5eda not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.550000 4733 scope.go:117] "RemoveContainer" containerID="afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.550183 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd\": container with ID starting with afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd not found: ID does not exist" containerID="afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.550213 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd"} err="failed to get container status \"afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd\": rpc error: code = NotFound desc = could not find container \"afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd\": container with ID starting with afc049fbf83728e7337456441c5fbda07ed30e2b7dd73e13098055e22e7188cd not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.550230 4733 scope.go:117] "RemoveContainer" containerID="aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.550430 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca\": container with ID starting with aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca not found: ID does not exist" containerID="aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.550463 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca"} err="failed to get container status \"aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca\": rpc error: code = NotFound desc = could not find container \"aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca\": container with ID starting with aa996c6972f45d4d5c1b6376c7be5089996244ab7ae1139f1e0b8e76c79089ca not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.550486 4733 scope.go:117] "RemoveContainer" containerID="a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.550686 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87\": container with ID starting with a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87 not found: ID does not exist" containerID="a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.550706 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87"} err="failed to get container status \"a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87\": rpc error: code = NotFound desc = could not find container \"a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87\": container with ID starting with a005804e1886a2cc1e0db35288fbe559cbee7ac9f4ff6c21b7005411a992cf87 not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.550718 4733 scope.go:117] "RemoveContainer" containerID="e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.550970 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95\": container with ID starting with e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95 not found: ID does not exist" containerID="e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.550996 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95"} err="failed to get container status \"e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95\": rpc error: code = NotFound desc = could not find container \"e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95\": container with ID starting with e01b15def4490527e145642eff53ac5b158c7b1d1f82b70e30e15e4c3b418a95 not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551012 4733 scope.go:117] "RemoveContainer" containerID="e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.551200 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc\": container with ID starting with e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc not found: ID does not exist" containerID="e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551223 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc"} err="failed to get container status \"e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc\": rpc error: code = NotFound desc = could not find container \"e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc\": container with ID starting with e7ff86b560df5d307e391e4a2f610074117a2728b35f3cbb2d2fe7817a34d2bc not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551247 4733 scope.go:117] "RemoveContainer" containerID="13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.551412 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018\": container with ID starting with 13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018 not found: ID does not exist" containerID="13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551430 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018"} err="failed to get container status \"13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018\": rpc error: code = NotFound desc = could not find container \"13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018\": container with ID starting with 13e541999baedad79919487bae37b66c3c5ba3068d8319233a8afec017e9e018 not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551442 4733 scope.go:117] "RemoveContainer" containerID="cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.551632 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d\": container with ID starting with cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d not found: ID does not exist" containerID="cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551653 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d"} err="failed to get container status \"cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d\": rpc error: code = NotFound desc = could not find container \"cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d\": container with ID starting with cd17a76031a3a2bf1779a2724807ccd64f82b9e3100716b0659a58c71deae51d not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551667 4733 scope.go:117] "RemoveContainer" containerID="da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.551897 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04\": container with ID starting with da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04 not found: ID does not exist" containerID="da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551923 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04"} err="failed to get container status \"da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04\": rpc error: code = NotFound desc = could not find container \"da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04\": container with ID starting with da343878da62a021822f0a84f1a2c5ed675cdc728a86fcfb641816a22cc34c04 not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.551939 4733 scope.go:117] "RemoveContainer" containerID="1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.552155 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0\": container with ID starting with 1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0 not found: ID does not exist" containerID="1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.552183 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0"} err="failed to get container status \"1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0\": rpc error: code = NotFound desc = could not find container \"1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0\": container with ID starting with 1b366b7ccb22a2d94ccb2c2f2734b532d592c76d284122afe5b9b0f3e41e77c0 not found: ID does not exist" Dec 04 18:04:12 crc kubenswrapper[4733]: I1204 18:04:12.647377 4733 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod3accb1a1-6ad9-42b9-9894-ba7b9e01d919"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod3accb1a1-6ad9-42b9-9894-ba7b9e01d919] : Timed out while waiting for systemd to remove kubepods-besteffort-pod3accb1a1_6ad9_42b9_9894_ba7b9e01d919.slice" Dec 04 18:04:12 crc kubenswrapper[4733]: E1204 18:04:12.647434 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod3accb1a1-6ad9-42b9-9894-ba7b9e01d919] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod3accb1a1-6ad9-42b9-9894-ba7b9e01d919] : Timed out while waiting for systemd to remove kubepods-besteffort-pod3accb1a1_6ad9_42b9_9894_ba7b9e01d919.slice" pod="openstack/ovn-controller-j58xt" podUID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" Dec 04 18:04:13 crc kubenswrapper[4733]: I1204 18:04:13.185532 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-j58xt" Dec 04 18:04:13 crc kubenswrapper[4733]: I1204 18:04:13.202762 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-j58xt"] Dec 04 18:04:13 crc kubenswrapper[4733]: I1204 18:04:13.209178 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-j58xt"] Dec 04 18:04:14 crc kubenswrapper[4733]: I1204 18:04:14.345449 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" path="/var/lib/kubelet/pods/3accb1a1-6ad9-42b9-9894-ba7b9e01d919/volumes" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.069256 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.163458 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78jw9\" (UniqueName: \"kubernetes.io/projected/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-kube-api-access-78jw9\") pod \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\" (UID: \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\") " Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.163592 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts\") pod \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\" (UID: \"08cdf00b-cf60-4e3a-9283-833e3a8d95bf\") " Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.164289 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "08cdf00b-cf60-4e3a-9283-833e3a8d95bf" (UID: "08cdf00b-cf60-4e3a-9283-833e3a8d95bf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.170773 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-kube-api-access-78jw9" (OuterVolumeSpecName: "kube-api-access-78jw9") pod "08cdf00b-cf60-4e3a-9283-833e3a8d95bf" (UID: "08cdf00b-cf60-4e3a-9283-833e3a8d95bf"). InnerVolumeSpecName "kube-api-access-78jw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.215212 4733 generic.go:334] "Generic (PLEG): container finished" podID="08cdf00b-cf60-4e3a-9283-833e3a8d95bf" containerID="ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a" exitCode=137 Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.215266 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementee67-account-delete-sxlxx" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.215292 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementee67-account-delete-sxlxx" event={"ID":"08cdf00b-cf60-4e3a-9283-833e3a8d95bf","Type":"ContainerDied","Data":"ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a"} Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.215330 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementee67-account-delete-sxlxx" event={"ID":"08cdf00b-cf60-4e3a-9283-833e3a8d95bf","Type":"ContainerDied","Data":"358da79aa4080332b38fbb6621db2dfd0b2f29bed19e6891cfa753d0cf58ca24"} Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.215371 4733 scope.go:117] "RemoveContainer" containerID="ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.241894 4733 scope.go:117] "RemoveContainer" containerID="ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a" Dec 04 18:04:16 crc kubenswrapper[4733]: E1204 18:04:16.242297 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a\": container with ID starting with ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a not found: ID does not exist" containerID="ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.242338 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a"} err="failed to get container status \"ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a\": rpc error: code = NotFound desc = could not find container \"ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a\": container with ID starting with ecf110464ca98d8347ea88527b864c959101704115aaa900deaa3fcb0402b45a not found: ID does not exist" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.256878 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementee67-account-delete-sxlxx"] Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.264790 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78jw9\" (UniqueName: \"kubernetes.io/projected/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-kube-api-access-78jw9\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.264827 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08cdf00b-cf60-4e3a-9283-833e3a8d95bf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.265576 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementee67-account-delete-sxlxx"] Dec 04 18:04:16 crc kubenswrapper[4733]: I1204 18:04:16.351994 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08cdf00b-cf60-4e3a-9283-833e3a8d95bf" path="/var/lib/kubelet/pods/08cdf00b-cf60-4e3a-9283-833e3a8d95bf/volumes" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.092506 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.184648 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw9hg\" (UniqueName: \"kubernetes.io/projected/3f1c1320-5093-464a-a9a0-8a4cb668d70f-kube-api-access-pw9hg\") pod \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\" (UID: \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\") " Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.184785 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts\") pod \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\" (UID: \"3f1c1320-5093-464a-a9a0-8a4cb668d70f\") " Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.185239 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f1c1320-5093-464a-a9a0-8a4cb668d70f" (UID: "3f1c1320-5093-464a-a9a0-8a4cb668d70f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.190389 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f1c1320-5093-464a-a9a0-8a4cb668d70f-kube-api-access-pw9hg" (OuterVolumeSpecName: "kube-api-access-pw9hg") pod "3f1c1320-5093-464a-a9a0-8a4cb668d70f" (UID: "3f1c1320-5093-464a-a9a0-8a4cb668d70f"). InnerVolumeSpecName "kube-api-access-pw9hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.224613 4733 generic.go:334] "Generic (PLEG): container finished" podID="3f1c1320-5093-464a-a9a0-8a4cb668d70f" containerID="1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414" exitCode=137 Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.224671 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder7119-account-delete-kzkjb" event={"ID":"3f1c1320-5093-464a-a9a0-8a4cb668d70f","Type":"ContainerDied","Data":"1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414"} Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.224706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder7119-account-delete-kzkjb" event={"ID":"3f1c1320-5093-464a-a9a0-8a4cb668d70f","Type":"ContainerDied","Data":"dd7e5dd3b4dd495bfcc89bbbb5553e1aa540defc806d3f337d11380f89e63395"} Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.224732 4733 scope.go:117] "RemoveContainer" containerID="1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.224878 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder7119-account-delete-kzkjb" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.247521 4733 scope.go:117] "RemoveContainer" containerID="1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414" Dec 04 18:04:17 crc kubenswrapper[4733]: E1204 18:04:17.247963 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414\": container with ID starting with 1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414 not found: ID does not exist" containerID="1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.248004 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414"} err="failed to get container status \"1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414\": rpc error: code = NotFound desc = could not find container \"1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414\": container with ID starting with 1921aa7bd062e05df09bcbfb3fb7e4b1c73db6ed8d3f51624a6cd87eba3ca414 not found: ID does not exist" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.269021 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder7119-account-delete-kzkjb"] Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.276095 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder7119-account-delete-kzkjb"] Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.285946 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw9hg\" (UniqueName: \"kubernetes.io/projected/3f1c1320-5093-464a-a9a0-8a4cb668d70f-kube-api-access-pw9hg\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:17 crc kubenswrapper[4733]: I1204 18:04:17.285977 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f1c1320-5093-464a-a9a0-8a4cb668d70f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:17 crc kubenswrapper[4733]: E1204 18:04:17.590105 4733 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 18:04:17 crc kubenswrapper[4733]: E1204 18:04:17.590188 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts podName:32b94636-cf95-42f9-8589-e38b550053af nodeName:}" failed. No retries permitted until 2025-12-04 18:04:49.590167996 +0000 UTC m=+1551.545529052 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts") pod "novacell06ac2-account-delete-rs267" (UID: "32b94636-cf95-42f9-8589-e38b550053af") : configmap "openstack-scripts" not found Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.186418 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.234390 4733 generic.go:334] "Generic (PLEG): container finished" podID="32b94636-cf95-42f9-8589-e38b550053af" containerID="ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7" exitCode=137 Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.234438 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell06ac2-account-delete-rs267" event={"ID":"32b94636-cf95-42f9-8589-e38b550053af","Type":"ContainerDied","Data":"ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7"} Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.234481 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell06ac2-account-delete-rs267" event={"ID":"32b94636-cf95-42f9-8589-e38b550053af","Type":"ContainerDied","Data":"9ac864fb7141d34ebf8b982d955f2c128644efd586c33ac7b0cdf84e93109fc7"} Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.234502 4733 scope.go:117] "RemoveContainer" containerID="ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.234612 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell06ac2-account-delete-rs267" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.253779 4733 scope.go:117] "RemoveContainer" containerID="ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7" Dec 04 18:04:18 crc kubenswrapper[4733]: E1204 18:04:18.254202 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7\": container with ID starting with ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7 not found: ID does not exist" containerID="ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.254248 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7"} err="failed to get container status \"ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7\": rpc error: code = NotFound desc = could not find container \"ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7\": container with ID starting with ffa93f0c672fa8d91f4de9a3d5cd98bd3dce467884d25b905eb3aee3fe577bb7 not found: ID does not exist" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.298877 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts\") pod \"32b94636-cf95-42f9-8589-e38b550053af\" (UID: \"32b94636-cf95-42f9-8589-e38b550053af\") " Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.299026 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hrxp\" (UniqueName: \"kubernetes.io/projected/32b94636-cf95-42f9-8589-e38b550053af-kube-api-access-4hrxp\") pod \"32b94636-cf95-42f9-8589-e38b550053af\" (UID: \"32b94636-cf95-42f9-8589-e38b550053af\") " Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.299602 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32b94636-cf95-42f9-8589-e38b550053af" (UID: "32b94636-cf95-42f9-8589-e38b550053af"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.311188 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b94636-cf95-42f9-8589-e38b550053af-kube-api-access-4hrxp" (OuterVolumeSpecName: "kube-api-access-4hrxp") pod "32b94636-cf95-42f9-8589-e38b550053af" (UID: "32b94636-cf95-42f9-8589-e38b550053af"). InnerVolumeSpecName "kube-api-access-4hrxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.356321 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f1c1320-5093-464a-a9a0-8a4cb668d70f" path="/var/lib/kubelet/pods/3f1c1320-5093-464a-a9a0-8a4cb668d70f/volumes" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.401604 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b94636-cf95-42f9-8589-e38b550053af-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.401686 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hrxp\" (UniqueName: \"kubernetes.io/projected/32b94636-cf95-42f9-8589-e38b550053af-kube-api-access-4hrxp\") on node \"crc\" DevicePath \"\"" Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.556604 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell06ac2-account-delete-rs267"] Dec 04 18:04:18 crc kubenswrapper[4733]: I1204 18:04:18.562406 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell06ac2-account-delete-rs267"] Dec 04 18:04:20 crc kubenswrapper[4733]: I1204 18:04:20.409964 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b94636-cf95-42f9-8589-e38b550053af" path="/var/lib/kubelet/pods/32b94636-cf95-42f9-8589-e38b550053af/volumes" Dec 04 18:05:15 crc kubenswrapper[4733]: I1204 18:05:15.362684 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:05:15 crc kubenswrapper[4733]: I1204 18:05:15.363406 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.006010 4733 scope.go:117] "RemoveContainer" containerID="e2276e988658554a414066aa0968e027716dad08111b838f90813477030b0565" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.043634 4733 scope.go:117] "RemoveContainer" containerID="1cf76a2cd64147788487fcf9fe8b7751a8a4db7c081716cc42c264a1a160bb4f" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.075130 4733 scope.go:117] "RemoveContainer" containerID="4c0f39dd57ee5768e8cb2fc129fedea35799ad4350c78f9e6bf0de97e2423216" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.101189 4733 scope.go:117] "RemoveContainer" containerID="04e97a72c5dc8120564a0e524dfcc8fc7aae822e190797b30bde855daa1981d9" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.128062 4733 scope.go:117] "RemoveContainer" containerID="da70a66a6d6be97c5ab3e3bb531110131ae53dcb8be21760ae2043a41665da7a" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.149234 4733 scope.go:117] "RemoveContainer" containerID="9ccae8be73d6c45906bab0a51e77d0e62fe11a561821293fc12bd3c1b094fd50" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.179880 4733 scope.go:117] "RemoveContainer" containerID="50af9429ee9c06ced96baa47e5ad56e3b3acb94ca909070661aefdb62d011c86" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.200233 4733 scope.go:117] "RemoveContainer" containerID="51718f0a98bc4f12648ae0c8e161095d9435571807a28e34ed23fb1ce1e33217" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.220981 4733 scope.go:117] "RemoveContainer" containerID="7c64cb40e986cfa71acbcbd9c5202683bf86fb0a870830d77be7d32b48479000" Dec 04 18:05:16 crc kubenswrapper[4733]: I1204 18:05:16.241061 4733 scope.go:117] "RemoveContainer" containerID="4a1a626a6d350a4c0d0cb45c40052b41650203413c0c302e9f5fe51aba7cd968" Dec 04 18:05:45 crc kubenswrapper[4733]: I1204 18:05:45.362413 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:05:45 crc kubenswrapper[4733]: I1204 18:05:45.363090 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.969496 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vgmxv"] Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970587 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970611 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-server" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970634 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="ceilometer-notification-agent" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970647 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="ceilometer-notification-agent" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970663 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32514f48-c09d-4bfc-a407-bc0b454ca946" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970678 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="32514f48-c09d-4bfc-a407-bc0b454ca946" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970699 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970712 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970734 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d28f771-2fba-42d9-8fc0-1107d17db527" containerName="memcached" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970749 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d28f771-2fba-42d9-8fc0-1107d17db527" containerName="memcached" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970778 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970792 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-api" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970839 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerName="rabbitmq" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970852 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerName="rabbitmq" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970878 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="swift-recon-cron" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970890 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="swift-recon-cron" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970916 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970930 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970955 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-metadata" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970968 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-metadata" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.970984 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerName="placement-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.970997 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerName="placement-api" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971016 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971030 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971053 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faf216c0-9cda-4f76-a4df-8855352db991" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971066 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="faf216c0-9cda-4f76-a4df-8855352db991" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971094 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerName="ovsdbserver-nb" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971154 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerName="ovsdbserver-nb" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971171 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971185 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971211 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d98e38-ee17-4974-94e7-6495564940be" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971225 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d98e38-ee17-4974-94e7-6495564940be" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971239 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="839d7537-ff12-4ee6-b36c-b192f130a6e4" containerName="nova-cell1-conductor-conductor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971252 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="839d7537-ff12-4ee6-b36c-b192f130a6e4" containerName="nova-cell1-conductor-conductor" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971268 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerName="proxy-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971281 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerName="proxy-server" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971298 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1c1320-5093-464a-a9a0-8a4cb668d70f" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971311 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1c1320-5093-464a-a9a0-8a4cb668d70f" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971338 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="sg-core" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971351 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="sg-core" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971375 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-updater" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971387 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-updater" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971408 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" containerName="init" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971420 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" containerName="init" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971442 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="proxy-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971454 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="proxy-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971467 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerName="placement-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971480 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerName="placement-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971505 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971519 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971536 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerName="probe" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971550 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerName="probe" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971573 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b94636-cf95-42f9-8589-e38b550053af" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971587 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b94636-cf95-42f9-8589-e38b550053af" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971606 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="ovsdbserver-sb" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971620 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="ovsdbserver-sb" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971643 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerName="rabbitmq" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971655 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerName="rabbitmq" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971673 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971687 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971705 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" containerName="keystone-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971720 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" containerName="keystone-api" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971779 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-updater" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971821 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-updater" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971913 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-reaper" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971962 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-reaper" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.971981 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75c5521d-9c39-4fa7-88ad-11eb375c6d9a" containerName="nova-scheduler-scheduler" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.971995 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75c5521d-9c39-4fa7-88ad-11eb375c6d9a" containerName="nova-scheduler-scheduler" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972020 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972033 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-server" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972051 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerName="glance-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972064 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerName="glance-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972087 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972100 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972116 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972129 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-server" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972147 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerName="neutron-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972270 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerName="neutron-api" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972287 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972300 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972319 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" containerName="mysql-bootstrap" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972332 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" containerName="mysql-bootstrap" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972380 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972429 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972449 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerName="galera" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972462 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerName="galera" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972480 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerName="setup-container" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972493 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerName="setup-container" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972514 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b77d7a1-e0fe-466e-9629-ac6b5bab61dc" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972531 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b77d7a1-e0fe-466e-9629-ac6b5bab61dc" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972567 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerName="barbican-keystone-listener-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972586 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerName="barbican-keystone-listener-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972606 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eecf779-672e-4469-a134-9ecb0ef1bdde" containerName="nova-cell0-conductor-conductor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972620 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eecf779-672e-4469-a134-9ecb0ef1bdde" containerName="nova-cell0-conductor-conductor" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972645 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerName="mysql-bootstrap" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972662 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerName="mysql-bootstrap" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972686 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerName="barbican-worker" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972699 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerName="barbican-worker" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972712 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerName="proxy-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972726 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerName="proxy-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972750 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="ceilometer-central-agent" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972765 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="ceilometer-central-agent" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972787 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" containerName="ovn-controller" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972836 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" containerName="ovn-controller" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972853 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf38da41-b6c0-4551-9ce9-32146ebc2160" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972866 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf38da41-b6c0-4551-9ce9-32146ebc2160" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972880 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerName="barbican-keystone-listener" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972893 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerName="barbican-keystone-listener" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972920 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972937 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972958 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerName="neutron-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.972971 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerName="neutron-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.972991 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658fbbbc-cb7d-4795-be06-76a46b589943" containerName="kube-state-metrics" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.973065 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="658fbbbc-cb7d-4795-be06-76a46b589943" containerName="kube-state-metrics" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.973099 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" containerName="galera" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.973114 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" containerName="galera" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.973978 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4047083c-6418-4814-8b24-9407494d0303" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977049 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4047083c-6418-4814-8b24-9407494d0303" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977081 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" containerName="dnsmasq-dns" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977098 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" containerName="dnsmasq-dns" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977117 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977133 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977150 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server-init" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977163 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server-init" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977183 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerName="setup-container" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977196 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerName="setup-container" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977212 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-expirer" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977224 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-expirer" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977241 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977254 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977274 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977286 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977311 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977326 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977341 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977354 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977379 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977392 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977413 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08cdf00b-cf60-4e3a-9283-833e3a8d95bf" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977426 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="08cdf00b-cf60-4e3a-9283-833e3a8d95bf" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977445 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977458 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977480 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977493 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977510 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" containerName="ovn-northd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977524 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" containerName="ovn-northd" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977539 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerName="barbican-worker-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977552 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerName="barbican-worker-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977565 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerName="cinder-scheduler" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977577 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerName="cinder-scheduler" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977603 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977617 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977636 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerName="glance-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977649 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerName="glance-log" Dec 04 18:05:57 crc kubenswrapper[4733]: E1204 18:05:57.977671 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="rsync" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.977683 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="rsync" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978054 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="32514f48-c09d-4bfc-a407-bc0b454ca946" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978074 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="ceilometer-central-agent" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978090 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-updater" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978106 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerName="glance-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978129 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" containerName="ovn-northd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978145 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerName="placement-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978159 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4047083c-6418-4814-8b24-9407494d0303" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978171 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b77d7a1-e0fe-466e-9629-ac6b5bab61dc" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978191 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978214 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-reaper" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978231 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovs-vswitchd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978246 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-expirer" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978260 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="swift-recon-cron" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978284 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerName="probe" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978301 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="sg-core" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978317 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf011d5-1c33-4ec6-8b1f-d1216a46d066" containerName="ovsdb-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978331 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b94636-cf95-42f9-8589-e38b550053af" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978356 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978368 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978392 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="229fa5f0-e006-4c04-8476-23a7842d292b" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978407 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978432 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="ovsdbserver-sb" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978450 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerName="barbican-keystone-listener-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978474 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75c5521d-9c39-4fa7-88ad-11eb375c6d9a" containerName="nova-scheduler-scheduler" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978497 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978515 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="faf216c0-9cda-4f76-a4df-8855352db991" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978533 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9cb46d6-3619-4035-8431-0b8552c5e690" containerName="galera" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978550 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-updater" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978562 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="proxy-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978888 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e605c62-64b4-4417-80bb-bc3387881f7a" containerName="rabbitmq" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978926 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerName="proxy-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978949 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="658fbbbc-cb7d-4795-be06-76a46b589943" containerName="kube-state-metrics" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978965 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="90549281-a4f7-416e-92dd-cea0e94c9af7" containerName="ceilometer-notification-agent" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978982 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7446c8ed-a97b-41a1-ae84-8f1eed8b203b" containerName="dnsmasq-dns" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.978998 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5f9b26-6c76-4af7-a811-d7d763f74ed6" containerName="barbican-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979021 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="rsync" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979043 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979062 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7df4d1e-71e7-4c40-a3e4-27b484799d09" containerName="cinder-scheduler" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979082 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c8b0056-e347-4f24-aeea-20680c33bf06" containerName="glance-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979099 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="08cdf00b-cf60-4e3a-9283-833e3a8d95bf" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979119 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979137 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979154 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eecf779-672e-4469-a134-9ecb0ef1bdde" containerName="nova-cell0-conductor-conductor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979172 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae2e2ab-b6af-4d4d-8288-9f23050553a0" containerName="ovsdbserver-nb" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979190 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979204 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="container-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979225 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f1c1320-5093-464a-a9a0-8a4cb668d70f" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979249 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-auditor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979261 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-server" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979279 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae61093e-d566-48cb-acc2-dd88c4512b8e" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979293 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerName="neutron-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979308 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d98e38-ee17-4974-94e7-6495564940be" containerName="mariadb-account-delete" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979327 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerName="barbican-worker" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979349 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9d9f43-f063-47fc-83b9-213c6ab06ef4" containerName="neutron-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979365 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3accb1a1-6ad9-42b9-9894-ba7b9e01d919" containerName="ovn-controller" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979383 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf38da41-b6c0-4551-9ce9-32146ebc2160" containerName="openstack-network-exporter" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979403 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="account-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979419 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1de9dcb-fad2-485d-b59b-14646f9ddfdd" containerName="barbican-keystone-listener" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979437 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d78c8335-c2e4-4c49-8b77-98fec3cab751" containerName="proxy-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979453 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="839d7537-ff12-4ee6-b36c-b192f130a6e4" containerName="nova-cell1-conductor-conductor" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979471 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd72019f-8a36-43af-8d73-7c3e804c1baa" containerName="nova-metadata-metadata" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979490 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f707420-7a54-4f2c-ac75-38e09d8dfb61" containerName="nova-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979508 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f55e362-d59c-4269-92c3-d5ca014a2ef1" containerName="rabbitmq" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979524 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75da96fa-8079-442d-8993-de5658a23f91" containerName="object-replicator" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979542 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5833da06-817f-4451-be68-9a9a94e5ef64" containerName="galera" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979558 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979570 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f575fafd-64f0-4401-82f1-f4a6a33b132e" containerName="placement-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979587 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="87f72b89-cea6-4243-8ec6-e1264b4901c3" containerName="barbican-worker-log" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979606 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="98545dd5-cfa2-4c6c-97e9-6985a1d5a60d" containerName="glance-httpd" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979622 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="12b58db1-9574-4081-a3c1-2853f003fa8f" containerName="cinder-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979640 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d28f771-2fba-42d9-8fc0-1107d17db527" containerName="memcached" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.979661 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd2ad1b-c8c0-4bdb-b46c-4bfd8d3fe1fa" containerName="keystone-api" Dec 04 18:05:57 crc kubenswrapper[4733]: I1204 18:05:57.981442 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.002103 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vgmxv"] Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.084099 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-catalog-content\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.084338 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmwgr\" (UniqueName: \"kubernetes.io/projected/9d14d9e2-606c-46c2-8326-5794c29d0c32-kube-api-access-vmwgr\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.084448 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-utilities\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.186075 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmwgr\" (UniqueName: \"kubernetes.io/projected/9d14d9e2-606c-46c2-8326-5794c29d0c32-kube-api-access-vmwgr\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.186142 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-utilities\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.186236 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-catalog-content\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.187293 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-catalog-content\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.187570 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-utilities\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.206905 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmwgr\" (UniqueName: \"kubernetes.io/projected/9d14d9e2-606c-46c2-8326-5794c29d0c32-kube-api-access-vmwgr\") pod \"certified-operators-vgmxv\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.320356 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:05:58 crc kubenswrapper[4733]: I1204 18:05:58.623555 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vgmxv"] Dec 04 18:05:59 crc kubenswrapper[4733]: I1204 18:05:59.565391 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerID="fd22bfd2e9b52af5a5a65011560ce7c1fafa380fb89acfc999fdddeef9ebab34" exitCode=0 Dec 04 18:05:59 crc kubenswrapper[4733]: I1204 18:05:59.565455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vgmxv" event={"ID":"9d14d9e2-606c-46c2-8326-5794c29d0c32","Type":"ContainerDied","Data":"fd22bfd2e9b52af5a5a65011560ce7c1fafa380fb89acfc999fdddeef9ebab34"} Dec 04 18:05:59 crc kubenswrapper[4733]: I1204 18:05:59.565889 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vgmxv" event={"ID":"9d14d9e2-606c-46c2-8326-5794c29d0c32","Type":"ContainerStarted","Data":"12801190410ee0d93f5c083c271302f3d0662a0e95c2dcb4253b4ad88844e24b"} Dec 04 18:06:04 crc kubenswrapper[4733]: I1204 18:06:04.610646 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vgmxv" event={"ID":"9d14d9e2-606c-46c2-8326-5794c29d0c32","Type":"ContainerStarted","Data":"c66678c24e05da5f46aa54210b20c5576a61314666b6c82fce300f09a42fc49f"} Dec 04 18:06:05 crc kubenswrapper[4733]: I1204 18:06:05.628779 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerID="c66678c24e05da5f46aa54210b20c5576a61314666b6c82fce300f09a42fc49f" exitCode=0 Dec 04 18:06:05 crc kubenswrapper[4733]: I1204 18:06:05.628837 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vgmxv" event={"ID":"9d14d9e2-606c-46c2-8326-5794c29d0c32","Type":"ContainerDied","Data":"c66678c24e05da5f46aa54210b20c5576a61314666b6c82fce300f09a42fc49f"} Dec 04 18:06:06 crc kubenswrapper[4733]: I1204 18:06:06.640961 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vgmxv" event={"ID":"9d14d9e2-606c-46c2-8326-5794c29d0c32","Type":"ContainerStarted","Data":"f3f567439952f0036242a59996a3223171a1ae3541639697fb8c5a0b8a5a6cab"} Dec 04 18:06:06 crc kubenswrapper[4733]: I1204 18:06:06.667310 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vgmxv" podStartSLOduration=2.8043624879999998 podStartE2EDuration="9.667279041s" podCreationTimestamp="2025-12-04 18:05:57 +0000 UTC" firstStartedPulling="2025-12-04 18:05:59.568835894 +0000 UTC m=+1621.524196940" lastFinishedPulling="2025-12-04 18:06:06.431752437 +0000 UTC m=+1628.387113493" observedRunningTime="2025-12-04 18:06:06.661444005 +0000 UTC m=+1628.616805061" watchObservedRunningTime="2025-12-04 18:06:06.667279041 +0000 UTC m=+1628.622640097" Dec 04 18:06:08 crc kubenswrapper[4733]: I1204 18:06:08.321502 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:06:08 crc kubenswrapper[4733]: I1204 18:06:08.321741 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:06:09 crc kubenswrapper[4733]: I1204 18:06:09.405075 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-vgmxv" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="registry-server" probeResult="failure" output=< Dec 04 18:06:09 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 18:06:09 crc kubenswrapper[4733]: > Dec 04 18:06:15 crc kubenswrapper[4733]: I1204 18:06:15.361905 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:06:15 crc kubenswrapper[4733]: I1204 18:06:15.362537 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:06:15 crc kubenswrapper[4733]: I1204 18:06:15.362621 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:06:15 crc kubenswrapper[4733]: I1204 18:06:15.363745 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:06:15 crc kubenswrapper[4733]: I1204 18:06:15.363907 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" gracePeriod=600 Dec 04 18:06:16 crc kubenswrapper[4733]: I1204 18:06:16.444566 4733 scope.go:117] "RemoveContainer" containerID="cf6ccb65296ed2b05152275d1569159933dbce9caea4d4b0634f3470dd6b2429" Dec 04 18:06:16 crc kubenswrapper[4733]: I1204 18:06:16.487317 4733 scope.go:117] "RemoveContainer" containerID="569e02b361ce9612c6faa846d0e150a84d2d650950ed49b4e0e824acc633aad4" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.100878 4733 scope.go:117] "RemoveContainer" containerID="45806a9fe665f29c9726301d1f3e4806eb4602973959b29e42f98d3c93819ee4" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.127715 4733 scope.go:117] "RemoveContainer" containerID="2335b1cf12fa0b82828e2f0e53c505bbc80375fc5b0a4726aee29ed9a4847793" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.166192 4733 scope.go:117] "RemoveContainer" containerID="77118deac3a99d7641547823e8365e3ec7939353fa68ed8de8ec7575b41dd783" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.192872 4733 scope.go:117] "RemoveContainer" containerID="8db2bb8eb14546ebea7175ee3d8b0c49509e28fb3250d5aa493e06d670d5943c" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.212167 4733 scope.go:117] "RemoveContainer" containerID="099a0dd66081b1742a35d04c3412573bb6ed9b1b5b14e8add80a8d0e79aa25ab" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.229556 4733 scope.go:117] "RemoveContainer" containerID="1471d44f86f056dac90145e65c91eb8e660b5f31b908dcb5f5b7cb79fc5a7064" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.250435 4733 scope.go:117] "RemoveContainer" containerID="ab9aabca56e540a2cedf02be5dda5c3c15082cf0c0d6c2af3a21c506250f6bcb" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.282997 4733 scope.go:117] "RemoveContainer" containerID="7daea43ad01541be1c71ecfc6cbf37ab9bcaf5bedff0f422646929d973829b99" Dec 04 18:06:17 crc kubenswrapper[4733]: E1204 18:06:17.314961 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.322237 4733 scope.go:117] "RemoveContainer" containerID="7c07b0d7d016fce0dca335c978df6940bc979f89cde5956c8726ce39d20debad" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.340628 4733 scope.go:117] "RemoveContainer" containerID="9fc2b7a8967966a8f3ff5322d84258451c9bdd69b827ca292e5e59c5d79425d0" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.357350 4733 scope.go:117] "RemoveContainer" containerID="643fab1fc6fa29e09ee7b9e662e869940d502d3a5ca9eea01011f6c047c63755" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.374722 4733 scope.go:117] "RemoveContainer" containerID="beeeaa11c7814632a3f71a9a39055495a734228f830097d6a2f2043823fae14a" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.390364 4733 scope.go:117] "RemoveContainer" containerID="e9f80fba5d806616b88bd80d77acee2d8113ec26a9b590174f91a0f9d6ab8ba9" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.414021 4733 scope.go:117] "RemoveContainer" containerID="f5e6f2c72112a2171e2ad78efad46bcf8b2b83073efb9334f19fa80660ebe091" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.431280 4733 scope.go:117] "RemoveContainer" containerID="d49942198858476882af324869180b206cbbcac041cbb1d3b56286f2de3f2d02" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.466157 4733 scope.go:117] "RemoveContainer" containerID="70c93568467ddeaf24343edeb7cc5662dba7fac3fe6fb7f124ae21b60667977d" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.753482 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" exitCode=0 Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.753555 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7"} Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.753615 4733 scope.go:117] "RemoveContainer" containerID="f1e31ad24991110558bf25e49f4521223e1cb18d3a3ee43198b8eff935ecc2a3" Dec 04 18:06:17 crc kubenswrapper[4733]: I1204 18:06:17.754572 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:06:17 crc kubenswrapper[4733]: E1204 18:06:17.755069 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:06:18 crc kubenswrapper[4733]: I1204 18:06:18.409666 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:06:18 crc kubenswrapper[4733]: I1204 18:06:18.490902 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:06:18 crc kubenswrapper[4733]: I1204 18:06:18.575644 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vgmxv"] Dec 04 18:06:18 crc kubenswrapper[4733]: I1204 18:06:18.649359 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-khxrh"] Dec 04 18:06:18 crc kubenswrapper[4733]: I1204 18:06:18.649581 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-khxrh" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerName="registry-server" containerID="cri-o://169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b" gracePeriod=2 Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.025449 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khxrh" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.053654 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vbxm\" (UniqueName: \"kubernetes.io/projected/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-kube-api-access-5vbxm\") pod \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.053783 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-catalog-content\") pod \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.053842 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-utilities\") pod \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\" (UID: \"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d\") " Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.054549 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-utilities" (OuterVolumeSpecName: "utilities") pod "9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" (UID: "9d0f24bd-3b66-4c30-9d2f-15ea83654f4d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.062088 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-kube-api-access-5vbxm" (OuterVolumeSpecName: "kube-api-access-5vbxm") pod "9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" (UID: "9d0f24bd-3b66-4c30-9d2f-15ea83654f4d"). InnerVolumeSpecName "kube-api-access-5vbxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.104012 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" (UID: "9d0f24bd-3b66-4c30-9d2f-15ea83654f4d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.154938 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.154968 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.154978 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vbxm\" (UniqueName: \"kubernetes.io/projected/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d-kube-api-access-5vbxm\") on node \"crc\" DevicePath \"\"" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.775331 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerID="169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b" exitCode=0 Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.775371 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khxrh" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.775397 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khxrh" event={"ID":"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d","Type":"ContainerDied","Data":"169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b"} Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.776525 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khxrh" event={"ID":"9d0f24bd-3b66-4c30-9d2f-15ea83654f4d","Type":"ContainerDied","Data":"5b870eaba2113352e722055c286690fc2345480266ab17b2f115cab9bf43e7a8"} Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.776547 4733 scope.go:117] "RemoveContainer" containerID="169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.795516 4733 scope.go:117] "RemoveContainer" containerID="8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.814729 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-khxrh"] Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.823703 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-khxrh"] Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.835156 4733 scope.go:117] "RemoveContainer" containerID="ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.872633 4733 scope.go:117] "RemoveContainer" containerID="169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b" Dec 04 18:06:19 crc kubenswrapper[4733]: E1204 18:06:19.873205 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b\": container with ID starting with 169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b not found: ID does not exist" containerID="169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.873237 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b"} err="failed to get container status \"169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b\": rpc error: code = NotFound desc = could not find container \"169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b\": container with ID starting with 169d835a8ca136a36d1d727768930c6c42c610b4512ff7ed13b1aa4f4ccc837b not found: ID does not exist" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.873262 4733 scope.go:117] "RemoveContainer" containerID="8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59" Dec 04 18:06:19 crc kubenswrapper[4733]: E1204 18:06:19.873638 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59\": container with ID starting with 8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59 not found: ID does not exist" containerID="8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.873683 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59"} err="failed to get container status \"8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59\": rpc error: code = NotFound desc = could not find container \"8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59\": container with ID starting with 8f1f096c8d291ca5c0d6f0309e325a1c8f9590d2e9e182b48d77574b4a516a59 not found: ID does not exist" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.873722 4733 scope.go:117] "RemoveContainer" containerID="ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da" Dec 04 18:06:19 crc kubenswrapper[4733]: E1204 18:06:19.874107 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da\": container with ID starting with ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da not found: ID does not exist" containerID="ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da" Dec 04 18:06:19 crc kubenswrapper[4733]: I1204 18:06:19.874130 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da"} err="failed to get container status \"ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da\": rpc error: code = NotFound desc = could not find container \"ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da\": container with ID starting with ac38741a01161d456e5674ce2dec24b231c45ed9d690d70a8df9871f051260da not found: ID does not exist" Dec 04 18:06:20 crc kubenswrapper[4733]: I1204 18:06:20.345645 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" path="/var/lib/kubelet/pods/9d0f24bd-3b66-4c30-9d2f-15ea83654f4d/volumes" Dec 04 18:06:28 crc kubenswrapper[4733]: I1204 18:06:28.344880 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:06:28 crc kubenswrapper[4733]: E1204 18:06:28.345946 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:06:41 crc kubenswrapper[4733]: I1204 18:06:41.335468 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:06:41 crc kubenswrapper[4733]: E1204 18:06:41.337563 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:06:56 crc kubenswrapper[4733]: I1204 18:06:56.336307 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:06:56 crc kubenswrapper[4733]: E1204 18:06:56.337092 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:07:10 crc kubenswrapper[4733]: I1204 18:07:10.335175 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:07:10 crc kubenswrapper[4733]: E1204 18:07:10.335942 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.737517 4733 scope.go:117] "RemoveContainer" containerID="d366e224d098b7e5af2be8bd08ebc779d71f9239fa811ef64c92bd18e26bace0" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.761183 4733 scope.go:117] "RemoveContainer" containerID="17effa55d5014bbb37b70198b0ff6c9472bb2d684386773a2411cee6b42323de" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.791384 4733 scope.go:117] "RemoveContainer" containerID="84b5297fb15dad792d9c940a5d4f0d67d6d13f94bd2b0ff3d1ae0a60960f7798" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.811588 4733 scope.go:117] "RemoveContainer" containerID="da28237687950c4a080dd5e279c384c4f78082957a7be8c3e467ac127824eb2b" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.841979 4733 scope.go:117] "RemoveContainer" containerID="ebae6d3dccdf5b42c07640f9dd249ddf0193a7662a64ce8f92ecd4ca8021c824" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.872715 4733 scope.go:117] "RemoveContainer" containerID="594133e5106de2f14d436d478b5b6fa76670e3bcd1bb710de01b002b98c8b79a" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.893492 4733 scope.go:117] "RemoveContainer" containerID="e20b5cfb354e7b5111ecacb69236c810870c0447a1bda743bdcf9810c1519d94" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.911529 4733 scope.go:117] "RemoveContainer" containerID="a6c4a59c75c37a61feea92a8e8114a64759adb31de0de0bbb5f0d5725e5b60bd" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.934288 4733 scope.go:117] "RemoveContainer" containerID="011962e29ba8da2577d347e04ff5bb5770e5fc4174508d0f00b388328c73e4f8" Dec 04 18:07:17 crc kubenswrapper[4733]: I1204 18:07:17.952560 4733 scope.go:117] "RemoveContainer" containerID="279356342a7696e8f66f5a4be3dcb02380255c870e5fc62c09803611ee661684" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.004660 4733 scope.go:117] "RemoveContainer" containerID="5cd076e1396593ece03d523ffdfe9f53cc72e35677e8aab21b5111c2bcce8fa1" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.028226 4733 scope.go:117] "RemoveContainer" containerID="03be397f987be20236ee4d9b4181dbb81db3d0b00ead6b8a03cfb6abed63df74" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.065787 4733 scope.go:117] "RemoveContainer" containerID="04ca6ac5aad6dfaafa91564d2deaf7c701fa148811b46f58283eb9c5dd6a73d0" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.096451 4733 scope.go:117] "RemoveContainer" containerID="9aa9ac8aec81ac8172b07681bf3447f72d402bc17fafe55ed65c2f79900cd3fb" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.126004 4733 scope.go:117] "RemoveContainer" containerID="9d57e43da79ab2c5d0b9a9159e21dc1548b4a16f3e0e5f81bcbb1076708a59e7" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.154076 4733 scope.go:117] "RemoveContainer" containerID="b4d8f3153bf28ff3a9f828d772e2a07d32adc19bd7e0e719a18409eed1f32ea2" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.174010 4733 scope.go:117] "RemoveContainer" containerID="3044b3de56fc15c0557352a235e182ec8eaffbf427385fd0dad2161e82e28cd2" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.194309 4733 scope.go:117] "RemoveContainer" containerID="a3eef607b3353284902508f0ced38edd93dbee9370ecc514d4db535fa52597df" Dec 04 18:07:18 crc kubenswrapper[4733]: I1204 18:07:18.222749 4733 scope.go:117] "RemoveContainer" containerID="e672d46632a950c5a99212c64ced6adecc83268c0542e4914c98073746bc1f0d" Dec 04 18:07:21 crc kubenswrapper[4733]: I1204 18:07:21.335099 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:07:21 crc kubenswrapper[4733]: E1204 18:07:21.336096 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:07:33 crc kubenswrapper[4733]: I1204 18:07:33.336339 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:07:33 crc kubenswrapper[4733]: E1204 18:07:33.337460 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:07:44 crc kubenswrapper[4733]: I1204 18:07:44.335613 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:07:44 crc kubenswrapper[4733]: E1204 18:07:44.336656 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:07:58 crc kubenswrapper[4733]: I1204 18:07:58.348509 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:07:58 crc kubenswrapper[4733]: E1204 18:07:58.349756 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:08:09 crc kubenswrapper[4733]: I1204 18:08:09.335525 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:08:09 crc kubenswrapper[4733]: E1204 18:08:09.336694 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:08:18 crc kubenswrapper[4733]: I1204 18:08:18.490354 4733 scope.go:117] "RemoveContainer" containerID="b5cab691f053ad9777f8ae8b5b15035ab043220183fb8752c859b471c466a3d8" Dec 04 18:08:18 crc kubenswrapper[4733]: I1204 18:08:18.548501 4733 scope.go:117] "RemoveContainer" containerID="6602dc5879168dcaaa99620ea72395421ca373d0f88940c5eada9244c55900a0" Dec 04 18:08:18 crc kubenswrapper[4733]: I1204 18:08:18.574745 4733 scope.go:117] "RemoveContainer" containerID="a663e1bc5727c5ca185760cb484948770d7432190c4cda6e4378bca5a497f3b1" Dec 04 18:08:18 crc kubenswrapper[4733]: I1204 18:08:18.606215 4733 scope.go:117] "RemoveContainer" containerID="92cb443d4f09ae45ef70d088a796c54e8df6db7a8ae4d5dd462a4b4e85ed0ade" Dec 04 18:08:18 crc kubenswrapper[4733]: I1204 18:08:18.647075 4733 scope.go:117] "RemoveContainer" containerID="9faff0680efdb357f0f7ab94276aac3340ec7dc913f22f91f8b0cae7b5683fd3" Dec 04 18:08:24 crc kubenswrapper[4733]: I1204 18:08:24.334968 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:08:24 crc kubenswrapper[4733]: E1204 18:08:24.336196 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:08:35 crc kubenswrapper[4733]: I1204 18:08:35.336752 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:08:35 crc kubenswrapper[4733]: E1204 18:08:35.337679 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:08:46 crc kubenswrapper[4733]: I1204 18:08:46.336780 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:08:46 crc kubenswrapper[4733]: E1204 18:08:46.337862 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.384632 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9vgw5"] Dec 04 18:08:56 crc kubenswrapper[4733]: E1204 18:08:56.385505 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerName="registry-server" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.385585 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerName="registry-server" Dec 04 18:08:56 crc kubenswrapper[4733]: E1204 18:08:56.385597 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerName="extract-content" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.385603 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerName="extract-content" Dec 04 18:08:56 crc kubenswrapper[4733]: E1204 18:08:56.385624 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerName="extract-utilities" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.385630 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerName="extract-utilities" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.385770 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d0f24bd-3b66-4c30-9d2f-15ea83654f4d" containerName="registry-server" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.386961 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.411947 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vgw5"] Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.530626 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-catalog-content\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.530685 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-utilities\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.530816 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl4f8\" (UniqueName: \"kubernetes.io/projected/04110d92-bb1e-4a1e-9648-530d499f075c-kube-api-access-tl4f8\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.632652 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-catalog-content\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.632713 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-utilities\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.632766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl4f8\" (UniqueName: \"kubernetes.io/projected/04110d92-bb1e-4a1e-9648-530d499f075c-kube-api-access-tl4f8\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.633234 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-catalog-content\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.633308 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-utilities\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.658147 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl4f8\" (UniqueName: \"kubernetes.io/projected/04110d92-bb1e-4a1e-9648-530d499f075c-kube-api-access-tl4f8\") pod \"redhat-marketplace-9vgw5\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:56 crc kubenswrapper[4733]: I1204 18:08:56.704129 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:08:57 crc kubenswrapper[4733]: I1204 18:08:57.169000 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vgw5"] Dec 04 18:08:57 crc kubenswrapper[4733]: I1204 18:08:57.283546 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vgw5" event={"ID":"04110d92-bb1e-4a1e-9648-530d499f075c","Type":"ContainerStarted","Data":"b5da5992c699ce0709ed517a4d003c91fdc8387f56ff37ad97b5bf0ed68acdf3"} Dec 04 18:08:57 crc kubenswrapper[4733]: I1204 18:08:57.336366 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:08:57 crc kubenswrapper[4733]: E1204 18:08:57.336552 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:08:58 crc kubenswrapper[4733]: I1204 18:08:58.295369 4733 generic.go:334] "Generic (PLEG): container finished" podID="04110d92-bb1e-4a1e-9648-530d499f075c" containerID="e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b" exitCode=0 Dec 04 18:08:58 crc kubenswrapper[4733]: I1204 18:08:58.295500 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vgw5" event={"ID":"04110d92-bb1e-4a1e-9648-530d499f075c","Type":"ContainerDied","Data":"e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b"} Dec 04 18:08:58 crc kubenswrapper[4733]: I1204 18:08:58.299295 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:09:00 crc kubenswrapper[4733]: I1204 18:09:00.316912 4733 generic.go:334] "Generic (PLEG): container finished" podID="04110d92-bb1e-4a1e-9648-530d499f075c" containerID="b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1" exitCode=0 Dec 04 18:09:00 crc kubenswrapper[4733]: I1204 18:09:00.317001 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vgw5" event={"ID":"04110d92-bb1e-4a1e-9648-530d499f075c","Type":"ContainerDied","Data":"b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1"} Dec 04 18:09:01 crc kubenswrapper[4733]: I1204 18:09:01.330051 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vgw5" event={"ID":"04110d92-bb1e-4a1e-9648-530d499f075c","Type":"ContainerStarted","Data":"2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de"} Dec 04 18:09:01 crc kubenswrapper[4733]: I1204 18:09:01.361388 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9vgw5" podStartSLOduration=2.920700059 podStartE2EDuration="5.361366558s" podCreationTimestamp="2025-12-04 18:08:56 +0000 UTC" firstStartedPulling="2025-12-04 18:08:58.298775438 +0000 UTC m=+1800.254136514" lastFinishedPulling="2025-12-04 18:09:00.739441927 +0000 UTC m=+1802.694803013" observedRunningTime="2025-12-04 18:09:01.356976391 +0000 UTC m=+1803.312337437" watchObservedRunningTime="2025-12-04 18:09:01.361366558 +0000 UTC m=+1803.316727644" Dec 04 18:09:06 crc kubenswrapper[4733]: I1204 18:09:06.705191 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:09:06 crc kubenswrapper[4733]: I1204 18:09:06.705535 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:09:06 crc kubenswrapper[4733]: I1204 18:09:06.786391 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:09:07 crc kubenswrapper[4733]: I1204 18:09:07.444436 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:09:10 crc kubenswrapper[4733]: I1204 18:09:10.582738 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vgw5"] Dec 04 18:09:10 crc kubenswrapper[4733]: I1204 18:09:10.583158 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9vgw5" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" containerName="registry-server" containerID="cri-o://2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de" gracePeriod=2 Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.003757 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.146889 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl4f8\" (UniqueName: \"kubernetes.io/projected/04110d92-bb1e-4a1e-9648-530d499f075c-kube-api-access-tl4f8\") pod \"04110d92-bb1e-4a1e-9648-530d499f075c\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.147132 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-catalog-content\") pod \"04110d92-bb1e-4a1e-9648-530d499f075c\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.147305 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-utilities\") pod \"04110d92-bb1e-4a1e-9648-530d499f075c\" (UID: \"04110d92-bb1e-4a1e-9648-530d499f075c\") " Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.149767 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-utilities" (OuterVolumeSpecName: "utilities") pod "04110d92-bb1e-4a1e-9648-530d499f075c" (UID: "04110d92-bb1e-4a1e-9648-530d499f075c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.157189 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04110d92-bb1e-4a1e-9648-530d499f075c-kube-api-access-tl4f8" (OuterVolumeSpecName: "kube-api-access-tl4f8") pod "04110d92-bb1e-4a1e-9648-530d499f075c" (UID: "04110d92-bb1e-4a1e-9648-530d499f075c"). InnerVolumeSpecName "kube-api-access-tl4f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.186973 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04110d92-bb1e-4a1e-9648-530d499f075c" (UID: "04110d92-bb1e-4a1e-9648-530d499f075c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.250839 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl4f8\" (UniqueName: \"kubernetes.io/projected/04110d92-bb1e-4a1e-9648-530d499f075c-kube-api-access-tl4f8\") on node \"crc\" DevicePath \"\"" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.250877 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.250890 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04110d92-bb1e-4a1e-9648-530d499f075c-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.336783 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:09:11 crc kubenswrapper[4733]: E1204 18:09:11.337316 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.418272 4733 generic.go:334] "Generic (PLEG): container finished" podID="04110d92-bb1e-4a1e-9648-530d499f075c" containerID="2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de" exitCode=0 Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.418309 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vgw5" event={"ID":"04110d92-bb1e-4a1e-9648-530d499f075c","Type":"ContainerDied","Data":"2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de"} Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.418338 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vgw5" event={"ID":"04110d92-bb1e-4a1e-9648-530d499f075c","Type":"ContainerDied","Data":"b5da5992c699ce0709ed517a4d003c91fdc8387f56ff37ad97b5bf0ed68acdf3"} Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.418353 4733 scope.go:117] "RemoveContainer" containerID="2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.418400 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vgw5" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.454924 4733 scope.go:117] "RemoveContainer" containerID="b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.458964 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vgw5"] Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.463717 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vgw5"] Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.493959 4733 scope.go:117] "RemoveContainer" containerID="e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.513608 4733 scope.go:117] "RemoveContainer" containerID="2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de" Dec 04 18:09:11 crc kubenswrapper[4733]: E1204 18:09:11.514067 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de\": container with ID starting with 2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de not found: ID does not exist" containerID="2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.514095 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de"} err="failed to get container status \"2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de\": rpc error: code = NotFound desc = could not find container \"2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de\": container with ID starting with 2a843c04ea80c23117a25289f845d14acbe0a0446244f7c94bec17f69cb391de not found: ID does not exist" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.514116 4733 scope.go:117] "RemoveContainer" containerID="b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1" Dec 04 18:09:11 crc kubenswrapper[4733]: E1204 18:09:11.514430 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1\": container with ID starting with b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1 not found: ID does not exist" containerID="b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.514448 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1"} err="failed to get container status \"b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1\": rpc error: code = NotFound desc = could not find container \"b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1\": container with ID starting with b681c6b9998976a2e22860b4a00f6784d75765b818dfbfff79e704b905f243d1 not found: ID does not exist" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.514462 4733 scope.go:117] "RemoveContainer" containerID="e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b" Dec 04 18:09:11 crc kubenswrapper[4733]: E1204 18:09:11.514696 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b\": container with ID starting with e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b not found: ID does not exist" containerID="e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b" Dec 04 18:09:11 crc kubenswrapper[4733]: I1204 18:09:11.514713 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b"} err="failed to get container status \"e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b\": rpc error: code = NotFound desc = could not find container \"e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b\": container with ID starting with e12f08993c025fc49e9300f3deceb6d9ff6e9e8325432e31e94ad882608fee2b not found: ID does not exist" Dec 04 18:09:12 crc kubenswrapper[4733]: I1204 18:09:12.350996 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" path="/var/lib/kubelet/pods/04110d92-bb1e-4a1e-9648-530d499f075c/volumes" Dec 04 18:09:18 crc kubenswrapper[4733]: I1204 18:09:18.794761 4733 scope.go:117] "RemoveContainer" containerID="9a63a5646f24f905ab643629b703d9fa8539f845df76e04dbee266c5247dbf9d" Dec 04 18:09:25 crc kubenswrapper[4733]: I1204 18:09:25.336506 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:09:25 crc kubenswrapper[4733]: E1204 18:09:25.337707 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:09:39 crc kubenswrapper[4733]: I1204 18:09:39.335439 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:09:39 crc kubenswrapper[4733]: E1204 18:09:39.336522 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:09:52 crc kubenswrapper[4733]: I1204 18:09:52.336852 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:09:52 crc kubenswrapper[4733]: E1204 18:09:52.338028 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:10:04 crc kubenswrapper[4733]: I1204 18:10:04.335481 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:10:04 crc kubenswrapper[4733]: E1204 18:10:04.336492 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:10:18 crc kubenswrapper[4733]: I1204 18:10:18.341209 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:10:18 crc kubenswrapper[4733]: E1204 18:10:18.341956 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:10:18 crc kubenswrapper[4733]: I1204 18:10:18.891846 4733 scope.go:117] "RemoveContainer" containerID="b97820459144423fb9dffe72aae0571a0723b467879078ea36fb82747da5d291" Dec 04 18:10:18 crc kubenswrapper[4733]: I1204 18:10:18.932285 4733 scope.go:117] "RemoveContainer" containerID="10830977b29b2d58716fd705aedb7bc5d4ff1a61db024a027be6c97602a43587" Dec 04 18:10:18 crc kubenswrapper[4733]: I1204 18:10:18.952288 4733 scope.go:117] "RemoveContainer" containerID="1aa9e2ec45e1b882430e2811102f36451e65ac3c8ef29ffaad0b50cf3519c1a3" Dec 04 18:10:30 crc kubenswrapper[4733]: I1204 18:10:30.336066 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:10:30 crc kubenswrapper[4733]: E1204 18:10:30.339092 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:10:45 crc kubenswrapper[4733]: I1204 18:10:45.335708 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:10:45 crc kubenswrapper[4733]: E1204 18:10:45.337082 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:10:58 crc kubenswrapper[4733]: I1204 18:10:58.342876 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:10:58 crc kubenswrapper[4733]: E1204 18:10:58.343902 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:11:12 crc kubenswrapper[4733]: I1204 18:11:12.336521 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:11:12 crc kubenswrapper[4733]: E1204 18:11:12.338661 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:11:25 crc kubenswrapper[4733]: I1204 18:11:25.336213 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:11:25 crc kubenswrapper[4733]: I1204 18:11:25.666415 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"247e7e47e31412293bb049df2983c101d961b202a8b6d949ad0f187702a52ddf"} Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.181488 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-972wg"] Dec 04 18:12:56 crc kubenswrapper[4733]: E1204 18:12:56.182693 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" containerName="extract-utilities" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.182720 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" containerName="extract-utilities" Dec 04 18:12:56 crc kubenswrapper[4733]: E1204 18:12:56.182758 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" containerName="extract-content" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.182771 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" containerName="extract-content" Dec 04 18:12:56 crc kubenswrapper[4733]: E1204 18:12:56.183662 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" containerName="registry-server" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.183980 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" containerName="registry-server" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.184297 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="04110d92-bb1e-4a1e-9648-530d499f075c" containerName="registry-server" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.188357 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.196245 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-972wg"] Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.262659 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-utilities\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.262728 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlkgv\" (UniqueName: \"kubernetes.io/projected/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-kube-api-access-nlkgv\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.262760 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-catalog-content\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.364827 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-utilities\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.364893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlkgv\" (UniqueName: \"kubernetes.io/projected/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-kube-api-access-nlkgv\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.365042 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-catalog-content\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.365730 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-catalog-content\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.365967 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-utilities\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.390050 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlkgv\" (UniqueName: \"kubernetes.io/projected/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-kube-api-access-nlkgv\") pod \"redhat-operators-972wg\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.519032 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:12:56 crc kubenswrapper[4733]: I1204 18:12:56.971299 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-972wg"] Dec 04 18:12:57 crc kubenswrapper[4733]: E1204 18:12:57.320539 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff0b823a_c1c2_42d3_800c_1d030cdc6adb.slice/crio-conmon-188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca.scope\": RecentStats: unable to find data in memory cache]" Dec 04 18:12:57 crc kubenswrapper[4733]: I1204 18:12:57.556180 4733 generic.go:334] "Generic (PLEG): container finished" podID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerID="188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca" exitCode=0 Dec 04 18:12:57 crc kubenswrapper[4733]: I1204 18:12:57.556263 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-972wg" event={"ID":"ff0b823a-c1c2-42d3-800c-1d030cdc6adb","Type":"ContainerDied","Data":"188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca"} Dec 04 18:12:57 crc kubenswrapper[4733]: I1204 18:12:57.556468 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-972wg" event={"ID":"ff0b823a-c1c2-42d3-800c-1d030cdc6adb","Type":"ContainerStarted","Data":"0fad2763f65374c3ab130fc43a838bd32b036d6a48cfa073254b69358d564bb1"} Dec 04 18:12:58 crc kubenswrapper[4733]: I1204 18:12:58.566733 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-972wg" event={"ID":"ff0b823a-c1c2-42d3-800c-1d030cdc6adb","Type":"ContainerStarted","Data":"89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660"} Dec 04 18:12:59 crc kubenswrapper[4733]: I1204 18:12:59.579441 4733 generic.go:334] "Generic (PLEG): container finished" podID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerID="89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660" exitCode=0 Dec 04 18:12:59 crc kubenswrapper[4733]: I1204 18:12:59.580526 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-972wg" event={"ID":"ff0b823a-c1c2-42d3-800c-1d030cdc6adb","Type":"ContainerDied","Data":"89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660"} Dec 04 18:13:00 crc kubenswrapper[4733]: I1204 18:13:00.589895 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-972wg" event={"ID":"ff0b823a-c1c2-42d3-800c-1d030cdc6adb","Type":"ContainerStarted","Data":"dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be"} Dec 04 18:13:00 crc kubenswrapper[4733]: I1204 18:13:00.618310 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-972wg" podStartSLOduration=2.176080978 podStartE2EDuration="4.618292929s" podCreationTimestamp="2025-12-04 18:12:56 +0000 UTC" firstStartedPulling="2025-12-04 18:12:57.557968474 +0000 UTC m=+2039.513329520" lastFinishedPulling="2025-12-04 18:13:00.000180395 +0000 UTC m=+2041.955541471" observedRunningTime="2025-12-04 18:13:00.616368149 +0000 UTC m=+2042.571729225" watchObservedRunningTime="2025-12-04 18:13:00.618292929 +0000 UTC m=+2042.573653975" Dec 04 18:13:06 crc kubenswrapper[4733]: I1204 18:13:06.519558 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:13:06 crc kubenswrapper[4733]: I1204 18:13:06.519990 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:13:07 crc kubenswrapper[4733]: I1204 18:13:07.594434 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-972wg" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="registry-server" probeResult="failure" output=< Dec 04 18:13:07 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 18:13:07 crc kubenswrapper[4733]: > Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.659685 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k4pm4"] Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.664718 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.675838 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4pm4"] Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.776421 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-utilities\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.776728 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnzb6\" (UniqueName: \"kubernetes.io/projected/79ede6b8-96d2-4872-bdb0-573af316391e-kube-api-access-gnzb6\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.776783 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-catalog-content\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.878166 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-utilities\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.878229 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnzb6\" (UniqueName: \"kubernetes.io/projected/79ede6b8-96d2-4872-bdb0-573af316391e-kube-api-access-gnzb6\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.878302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-catalog-content\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.878941 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-catalog-content\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.879056 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-utilities\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.899882 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnzb6\" (UniqueName: \"kubernetes.io/projected/79ede6b8-96d2-4872-bdb0-573af316391e-kube-api-access-gnzb6\") pod \"community-operators-k4pm4\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:09 crc kubenswrapper[4733]: I1204 18:13:09.996301 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:10 crc kubenswrapper[4733]: W1204 18:13:10.386924 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79ede6b8_96d2_4872_bdb0_573af316391e.slice/crio-85f5b0d59b442ef27484853e2df980fcf362f4cbab9495f0b2e426dd0f4c0824 WatchSource:0}: Error finding container 85f5b0d59b442ef27484853e2df980fcf362f4cbab9495f0b2e426dd0f4c0824: Status 404 returned error can't find the container with id 85f5b0d59b442ef27484853e2df980fcf362f4cbab9495f0b2e426dd0f4c0824 Dec 04 18:13:10 crc kubenswrapper[4733]: I1204 18:13:10.393471 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4pm4"] Dec 04 18:13:10 crc kubenswrapper[4733]: I1204 18:13:10.716569 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4pm4" event={"ID":"79ede6b8-96d2-4872-bdb0-573af316391e","Type":"ContainerStarted","Data":"85f5b0d59b442ef27484853e2df980fcf362f4cbab9495f0b2e426dd0f4c0824"} Dec 04 18:13:11 crc kubenswrapper[4733]: I1204 18:13:11.729018 4733 generic.go:334] "Generic (PLEG): container finished" podID="79ede6b8-96d2-4872-bdb0-573af316391e" containerID="f6838dae9725c94d4ba9205ac14da84db281dca7690ff8d1d248c08c30d91ec9" exitCode=0 Dec 04 18:13:11 crc kubenswrapper[4733]: I1204 18:13:11.729095 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4pm4" event={"ID":"79ede6b8-96d2-4872-bdb0-573af316391e","Type":"ContainerDied","Data":"f6838dae9725c94d4ba9205ac14da84db281dca7690ff8d1d248c08c30d91ec9"} Dec 04 18:13:12 crc kubenswrapper[4733]: I1204 18:13:12.740494 4733 generic.go:334] "Generic (PLEG): container finished" podID="79ede6b8-96d2-4872-bdb0-573af316391e" containerID="5bdca859ded93a4b4302771eefd90c01607de3634c4023a179b2c453a9fa862b" exitCode=0 Dec 04 18:13:12 crc kubenswrapper[4733]: I1204 18:13:12.740622 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4pm4" event={"ID":"79ede6b8-96d2-4872-bdb0-573af316391e","Type":"ContainerDied","Data":"5bdca859ded93a4b4302771eefd90c01607de3634c4023a179b2c453a9fa862b"} Dec 04 18:13:13 crc kubenswrapper[4733]: I1204 18:13:13.754309 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4pm4" event={"ID":"79ede6b8-96d2-4872-bdb0-573af316391e","Type":"ContainerStarted","Data":"b1589c56e9b6e9d06d75d9946fb2883989bcadad43431e8e27754b10433f4ab4"} Dec 04 18:13:13 crc kubenswrapper[4733]: I1204 18:13:13.785087 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k4pm4" podStartSLOduration=3.356328728 podStartE2EDuration="4.78504719s" podCreationTimestamp="2025-12-04 18:13:09 +0000 UTC" firstStartedPulling="2025-12-04 18:13:11.731004941 +0000 UTC m=+2053.686365987" lastFinishedPulling="2025-12-04 18:13:13.159723363 +0000 UTC m=+2055.115084449" observedRunningTime="2025-12-04 18:13:13.782426101 +0000 UTC m=+2055.737787147" watchObservedRunningTime="2025-12-04 18:13:13.78504719 +0000 UTC m=+2055.740408266" Dec 04 18:13:16 crc kubenswrapper[4733]: I1204 18:13:16.594872 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:13:16 crc kubenswrapper[4733]: I1204 18:13:16.656100 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:13:17 crc kubenswrapper[4733]: I1204 18:13:17.653925 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-972wg"] Dec 04 18:13:17 crc kubenswrapper[4733]: I1204 18:13:17.806501 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-972wg" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="registry-server" containerID="cri-o://dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be" gracePeriod=2 Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.779653 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.843564 4733 generic.go:334] "Generic (PLEG): container finished" podID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerID="dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be" exitCode=0 Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.843616 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-972wg" event={"ID":"ff0b823a-c1c2-42d3-800c-1d030cdc6adb","Type":"ContainerDied","Data":"dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be"} Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.843653 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-972wg" event={"ID":"ff0b823a-c1c2-42d3-800c-1d030cdc6adb","Type":"ContainerDied","Data":"0fad2763f65374c3ab130fc43a838bd32b036d6a48cfa073254b69358d564bb1"} Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.843661 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-972wg" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.843673 4733 scope.go:117] "RemoveContainer" containerID="dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.862361 4733 scope.go:117] "RemoveContainer" containerID="89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.878416 4733 scope.go:117] "RemoveContainer" containerID="188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.904429 4733 scope.go:117] "RemoveContainer" containerID="dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be" Dec 04 18:13:18 crc kubenswrapper[4733]: E1204 18:13:18.904922 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be\": container with ID starting with dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be not found: ID does not exist" containerID="dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.904966 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be"} err="failed to get container status \"dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be\": rpc error: code = NotFound desc = could not find container \"dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be\": container with ID starting with dd8f7c7713cb948eddac9aa945c053d3499f3476274923a5bb4094242bc591be not found: ID does not exist" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.904992 4733 scope.go:117] "RemoveContainer" containerID="89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660" Dec 04 18:13:18 crc kubenswrapper[4733]: E1204 18:13:18.905398 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660\": container with ID starting with 89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660 not found: ID does not exist" containerID="89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.905449 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660"} err="failed to get container status \"89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660\": rpc error: code = NotFound desc = could not find container \"89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660\": container with ID starting with 89bbb9f93aa49f46c71bf4d2c493195de648f89aec01c07c8d8d853b8f596660 not found: ID does not exist" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.905483 4733 scope.go:117] "RemoveContainer" containerID="188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca" Dec 04 18:13:18 crc kubenswrapper[4733]: E1204 18:13:18.905813 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca\": container with ID starting with 188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca not found: ID does not exist" containerID="188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.905835 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca"} err="failed to get container status \"188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca\": rpc error: code = NotFound desc = could not find container \"188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca\": container with ID starting with 188fccc9397fd07661a6865e533553062961be2ab5ed76b86b78c40b23a040ca not found: ID does not exist" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.942318 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-utilities\") pod \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.942422 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-catalog-content\") pod \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.942532 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlkgv\" (UniqueName: \"kubernetes.io/projected/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-kube-api-access-nlkgv\") pod \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\" (UID: \"ff0b823a-c1c2-42d3-800c-1d030cdc6adb\") " Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.943337 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-utilities" (OuterVolumeSpecName: "utilities") pod "ff0b823a-c1c2-42d3-800c-1d030cdc6adb" (UID: "ff0b823a-c1c2-42d3-800c-1d030cdc6adb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:13:18 crc kubenswrapper[4733]: I1204 18:13:18.947731 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-kube-api-access-nlkgv" (OuterVolumeSpecName: "kube-api-access-nlkgv") pod "ff0b823a-c1c2-42d3-800c-1d030cdc6adb" (UID: "ff0b823a-c1c2-42d3-800c-1d030cdc6adb"). InnerVolumeSpecName "kube-api-access-nlkgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:13:19 crc kubenswrapper[4733]: I1204 18:13:19.044250 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlkgv\" (UniqueName: \"kubernetes.io/projected/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-kube-api-access-nlkgv\") on node \"crc\" DevicePath \"\"" Dec 04 18:13:19 crc kubenswrapper[4733]: I1204 18:13:19.044291 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:13:19 crc kubenswrapper[4733]: I1204 18:13:19.089238 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff0b823a-c1c2-42d3-800c-1d030cdc6adb" (UID: "ff0b823a-c1c2-42d3-800c-1d030cdc6adb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:13:19 crc kubenswrapper[4733]: I1204 18:13:19.145313 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0b823a-c1c2-42d3-800c-1d030cdc6adb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:13:19 crc kubenswrapper[4733]: I1204 18:13:19.180219 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-972wg"] Dec 04 18:13:19 crc kubenswrapper[4733]: I1204 18:13:19.185301 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-972wg"] Dec 04 18:13:19 crc kubenswrapper[4733]: I1204 18:13:19.997075 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:19 crc kubenswrapper[4733]: I1204 18:13:19.997329 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:20 crc kubenswrapper[4733]: I1204 18:13:20.056445 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:20 crc kubenswrapper[4733]: I1204 18:13:20.357039 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" path="/var/lib/kubelet/pods/ff0b823a-c1c2-42d3-800c-1d030cdc6adb/volumes" Dec 04 18:13:20 crc kubenswrapper[4733]: I1204 18:13:20.940560 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:22 crc kubenswrapper[4733]: I1204 18:13:22.049421 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k4pm4"] Dec 04 18:13:23 crc kubenswrapper[4733]: I1204 18:13:23.886010 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k4pm4" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" containerName="registry-server" containerID="cri-o://b1589c56e9b6e9d06d75d9946fb2883989bcadad43431e8e27754b10433f4ab4" gracePeriod=2 Dec 04 18:13:24 crc kubenswrapper[4733]: I1204 18:13:24.900788 4733 generic.go:334] "Generic (PLEG): container finished" podID="79ede6b8-96d2-4872-bdb0-573af316391e" containerID="b1589c56e9b6e9d06d75d9946fb2883989bcadad43431e8e27754b10433f4ab4" exitCode=0 Dec 04 18:13:24 crc kubenswrapper[4733]: I1204 18:13:24.900888 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4pm4" event={"ID":"79ede6b8-96d2-4872-bdb0-573af316391e","Type":"ContainerDied","Data":"b1589c56e9b6e9d06d75d9946fb2883989bcadad43431e8e27754b10433f4ab4"} Dec 04 18:13:24 crc kubenswrapper[4733]: I1204 18:13:24.901221 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4pm4" event={"ID":"79ede6b8-96d2-4872-bdb0-573af316391e","Type":"ContainerDied","Data":"85f5b0d59b442ef27484853e2df980fcf362f4cbab9495f0b2e426dd0f4c0824"} Dec 04 18:13:24 crc kubenswrapper[4733]: I1204 18:13:24.901243 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85f5b0d59b442ef27484853e2df980fcf362f4cbab9495f0b2e426dd0f4c0824" Dec 04 18:13:24 crc kubenswrapper[4733]: I1204 18:13:24.900948 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.046521 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-catalog-content\") pod \"79ede6b8-96d2-4872-bdb0-573af316391e\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.046602 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnzb6\" (UniqueName: \"kubernetes.io/projected/79ede6b8-96d2-4872-bdb0-573af316391e-kube-api-access-gnzb6\") pod \"79ede6b8-96d2-4872-bdb0-573af316391e\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.046742 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-utilities\") pod \"79ede6b8-96d2-4872-bdb0-573af316391e\" (UID: \"79ede6b8-96d2-4872-bdb0-573af316391e\") " Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.048309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-utilities" (OuterVolumeSpecName: "utilities") pod "79ede6b8-96d2-4872-bdb0-573af316391e" (UID: "79ede6b8-96d2-4872-bdb0-573af316391e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.056084 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ede6b8-96d2-4872-bdb0-573af316391e-kube-api-access-gnzb6" (OuterVolumeSpecName: "kube-api-access-gnzb6") pod "79ede6b8-96d2-4872-bdb0-573af316391e" (UID: "79ede6b8-96d2-4872-bdb0-573af316391e"). InnerVolumeSpecName "kube-api-access-gnzb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.124188 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79ede6b8-96d2-4872-bdb0-573af316391e" (UID: "79ede6b8-96d2-4872-bdb0-573af316391e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.148904 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.148939 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79ede6b8-96d2-4872-bdb0-573af316391e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.148953 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnzb6\" (UniqueName: \"kubernetes.io/projected/79ede6b8-96d2-4872-bdb0-573af316391e-kube-api-access-gnzb6\") on node \"crc\" DevicePath \"\"" Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.909644 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4pm4" Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.952839 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k4pm4"] Dec 04 18:13:25 crc kubenswrapper[4733]: I1204 18:13:25.959200 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k4pm4"] Dec 04 18:13:26 crc kubenswrapper[4733]: I1204 18:13:26.353464 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" path="/var/lib/kubelet/pods/79ede6b8-96d2-4872-bdb0-573af316391e/volumes" Dec 04 18:13:45 crc kubenswrapper[4733]: I1204 18:13:45.361478 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:13:45 crc kubenswrapper[4733]: I1204 18:13:45.362350 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:14:15 crc kubenswrapper[4733]: I1204 18:14:15.361574 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:14:15 crc kubenswrapper[4733]: I1204 18:14:15.362368 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:14:45 crc kubenswrapper[4733]: I1204 18:14:45.361941 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:14:45 crc kubenswrapper[4733]: I1204 18:14:45.362587 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:14:45 crc kubenswrapper[4733]: I1204 18:14:45.362658 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:14:45 crc kubenswrapper[4733]: I1204 18:14:45.363543 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"247e7e47e31412293bb049df2983c101d961b202a8b6d949ad0f187702a52ddf"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:14:45 crc kubenswrapper[4733]: I1204 18:14:45.363647 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://247e7e47e31412293bb049df2983c101d961b202a8b6d949ad0f187702a52ddf" gracePeriod=600 Dec 04 18:14:45 crc kubenswrapper[4733]: I1204 18:14:45.659510 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="247e7e47e31412293bb049df2983c101d961b202a8b6d949ad0f187702a52ddf" exitCode=0 Dec 04 18:14:45 crc kubenswrapper[4733]: I1204 18:14:45.659550 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"247e7e47e31412293bb049df2983c101d961b202a8b6d949ad0f187702a52ddf"} Dec 04 18:14:45 crc kubenswrapper[4733]: I1204 18:14:45.659598 4733 scope.go:117] "RemoveContainer" containerID="7863d7712d0a66c69824995d7c961edbad5a44e70b8abec6282641f900e1cad7" Dec 04 18:14:46 crc kubenswrapper[4733]: I1204 18:14:46.672621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c"} Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.163166 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr"] Dec 04 18:15:00 crc kubenswrapper[4733]: E1204 18:15:00.164080 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" containerName="registry-server" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164099 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" containerName="registry-server" Dec 04 18:15:00 crc kubenswrapper[4733]: E1204 18:15:00.164123 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="registry-server" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164131 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="registry-server" Dec 04 18:15:00 crc kubenswrapper[4733]: E1204 18:15:00.164146 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" containerName="extract-utilities" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164154 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" containerName="extract-utilities" Dec 04 18:15:00 crc kubenswrapper[4733]: E1204 18:15:00.164170 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" containerName="extract-content" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164177 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" containerName="extract-content" Dec 04 18:15:00 crc kubenswrapper[4733]: E1204 18:15:00.164192 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="extract-utilities" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164200 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="extract-utilities" Dec 04 18:15:00 crc kubenswrapper[4733]: E1204 18:15:00.164218 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="extract-content" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164226 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="extract-content" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164389 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0b823a-c1c2-42d3-800c-1d030cdc6adb" containerName="registry-server" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164410 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ede6b8-96d2-4872-bdb0-573af316391e" containerName="registry-server" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.164945 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.167224 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.167733 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.178134 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr"] Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.299319 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44e4ddcf-12df-4fad-bb44-96feea2c342e-config-volume\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.299436 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44e4ddcf-12df-4fad-bb44-96feea2c342e-secret-volume\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.299480 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmq2n\" (UniqueName: \"kubernetes.io/projected/44e4ddcf-12df-4fad-bb44-96feea2c342e-kube-api-access-xmq2n\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.402149 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmq2n\" (UniqueName: \"kubernetes.io/projected/44e4ddcf-12df-4fad-bb44-96feea2c342e-kube-api-access-xmq2n\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.402325 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44e4ddcf-12df-4fad-bb44-96feea2c342e-config-volume\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.402463 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44e4ddcf-12df-4fad-bb44-96feea2c342e-secret-volume\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.403399 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44e4ddcf-12df-4fad-bb44-96feea2c342e-config-volume\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.410289 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44e4ddcf-12df-4fad-bb44-96feea2c342e-secret-volume\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.419225 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmq2n\" (UniqueName: \"kubernetes.io/projected/44e4ddcf-12df-4fad-bb44-96feea2c342e-kube-api-access-xmq2n\") pod \"collect-profiles-29414535-cxqsr\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.495023 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:00 crc kubenswrapper[4733]: I1204 18:15:00.938355 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr"] Dec 04 18:15:00 crc kubenswrapper[4733]: W1204 18:15:00.950540 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44e4ddcf_12df_4fad_bb44_96feea2c342e.slice/crio-c588e34e9665426183e95cbcac469f7439764763ab014b76441239e3c60a7a85 WatchSource:0}: Error finding container c588e34e9665426183e95cbcac469f7439764763ab014b76441239e3c60a7a85: Status 404 returned error can't find the container with id c588e34e9665426183e95cbcac469f7439764763ab014b76441239e3c60a7a85 Dec 04 18:15:01 crc kubenswrapper[4733]: I1204 18:15:01.803045 4733 generic.go:334] "Generic (PLEG): container finished" podID="44e4ddcf-12df-4fad-bb44-96feea2c342e" containerID="e3013392776988c35cf517b0fbbdfb7ad63bab4e41f8b5cb44dca286bf157f5b" exitCode=0 Dec 04 18:15:01 crc kubenswrapper[4733]: I1204 18:15:01.803115 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" event={"ID":"44e4ddcf-12df-4fad-bb44-96feea2c342e","Type":"ContainerDied","Data":"e3013392776988c35cf517b0fbbdfb7ad63bab4e41f8b5cb44dca286bf157f5b"} Dec 04 18:15:01 crc kubenswrapper[4733]: I1204 18:15:01.803432 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" event={"ID":"44e4ddcf-12df-4fad-bb44-96feea2c342e","Type":"ContainerStarted","Data":"c588e34e9665426183e95cbcac469f7439764763ab014b76441239e3c60a7a85"} Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.119607 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.241375 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44e4ddcf-12df-4fad-bb44-96feea2c342e-config-volume\") pod \"44e4ddcf-12df-4fad-bb44-96feea2c342e\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.241459 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmq2n\" (UniqueName: \"kubernetes.io/projected/44e4ddcf-12df-4fad-bb44-96feea2c342e-kube-api-access-xmq2n\") pod \"44e4ddcf-12df-4fad-bb44-96feea2c342e\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.241516 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44e4ddcf-12df-4fad-bb44-96feea2c342e-secret-volume\") pod \"44e4ddcf-12df-4fad-bb44-96feea2c342e\" (UID: \"44e4ddcf-12df-4fad-bb44-96feea2c342e\") " Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.242203 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44e4ddcf-12df-4fad-bb44-96feea2c342e-config-volume" (OuterVolumeSpecName: "config-volume") pod "44e4ddcf-12df-4fad-bb44-96feea2c342e" (UID: "44e4ddcf-12df-4fad-bb44-96feea2c342e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.247626 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44e4ddcf-12df-4fad-bb44-96feea2c342e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "44e4ddcf-12df-4fad-bb44-96feea2c342e" (UID: "44e4ddcf-12df-4fad-bb44-96feea2c342e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.248309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44e4ddcf-12df-4fad-bb44-96feea2c342e-kube-api-access-xmq2n" (OuterVolumeSpecName: "kube-api-access-xmq2n") pod "44e4ddcf-12df-4fad-bb44-96feea2c342e" (UID: "44e4ddcf-12df-4fad-bb44-96feea2c342e"). InnerVolumeSpecName "kube-api-access-xmq2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.342662 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44e4ddcf-12df-4fad-bb44-96feea2c342e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.342705 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmq2n\" (UniqueName: \"kubernetes.io/projected/44e4ddcf-12df-4fad-bb44-96feea2c342e-kube-api-access-xmq2n\") on node \"crc\" DevicePath \"\"" Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.342723 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44e4ddcf-12df-4fad-bb44-96feea2c342e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.823645 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" event={"ID":"44e4ddcf-12df-4fad-bb44-96feea2c342e","Type":"ContainerDied","Data":"c588e34e9665426183e95cbcac469f7439764763ab014b76441239e3c60a7a85"} Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.824018 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c588e34e9665426183e95cbcac469f7439764763ab014b76441239e3c60a7a85" Dec 04 18:15:03 crc kubenswrapper[4733]: I1204 18:15:03.823738 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr" Dec 04 18:15:04 crc kubenswrapper[4733]: I1204 18:15:04.217989 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68"] Dec 04 18:15:04 crc kubenswrapper[4733]: I1204 18:15:04.225939 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414490-dhv68"] Dec 04 18:15:04 crc kubenswrapper[4733]: I1204 18:15:04.349722 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79f2ad87-1b0d-40ea-8bf4-51fb26df839f" path="/var/lib/kubelet/pods/79f2ad87-1b0d-40ea-8bf4-51fb26df839f/volumes" Dec 04 18:15:19 crc kubenswrapper[4733]: I1204 18:15:19.114076 4733 scope.go:117] "RemoveContainer" containerID="d24abf46b2c44c997d119b503f9ba47bad76b3c36c274ef3e41aae3fd7aab77c" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.571222 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f7dh5"] Dec 04 18:16:29 crc kubenswrapper[4733]: E1204 18:16:29.572450 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44e4ddcf-12df-4fad-bb44-96feea2c342e" containerName="collect-profiles" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.572479 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="44e4ddcf-12df-4fad-bb44-96feea2c342e" containerName="collect-profiles" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.572776 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="44e4ddcf-12df-4fad-bb44-96feea2c342e" containerName="collect-profiles" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.579601 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.585376 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f7dh5"] Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.584613 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-utilities\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.594336 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2zdn\" (UniqueName: \"kubernetes.io/projected/e5009ea9-a807-4453-8f7b-73b987bc5aa7-kube-api-access-v2zdn\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.594472 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-catalog-content\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.696129 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-utilities\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.696178 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2zdn\" (UniqueName: \"kubernetes.io/projected/e5009ea9-a807-4453-8f7b-73b987bc5aa7-kube-api-access-v2zdn\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.696204 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-catalog-content\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.696756 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-catalog-content\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.697026 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-utilities\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.723107 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2zdn\" (UniqueName: \"kubernetes.io/projected/e5009ea9-a807-4453-8f7b-73b987bc5aa7-kube-api-access-v2zdn\") pod \"certified-operators-f7dh5\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:29 crc kubenswrapper[4733]: I1204 18:16:29.913446 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:30 crc kubenswrapper[4733]: I1204 18:16:30.420759 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f7dh5"] Dec 04 18:16:30 crc kubenswrapper[4733]: I1204 18:16:30.621623 4733 generic.go:334] "Generic (PLEG): container finished" podID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerID="5fb225c1eeb12ed62644709887ebe8f8dd4d6600d732030c2ef37fe40df1aaa8" exitCode=0 Dec 04 18:16:30 crc kubenswrapper[4733]: I1204 18:16:30.621707 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7dh5" event={"ID":"e5009ea9-a807-4453-8f7b-73b987bc5aa7","Type":"ContainerDied","Data":"5fb225c1eeb12ed62644709887ebe8f8dd4d6600d732030c2ef37fe40df1aaa8"} Dec 04 18:16:30 crc kubenswrapper[4733]: I1204 18:16:30.621777 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7dh5" event={"ID":"e5009ea9-a807-4453-8f7b-73b987bc5aa7","Type":"ContainerStarted","Data":"5ad22598bb8428fac98b2f9251c7eb9699b640307cab9535519929517fb9470a"} Dec 04 18:16:30 crc kubenswrapper[4733]: I1204 18:16:30.623562 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:16:31 crc kubenswrapper[4733]: I1204 18:16:31.632740 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7dh5" event={"ID":"e5009ea9-a807-4453-8f7b-73b987bc5aa7","Type":"ContainerStarted","Data":"71fb1401d2180c86d22bd0a9a8f99d4c63a5b4c9c9edba91e55011192978304e"} Dec 04 18:16:32 crc kubenswrapper[4733]: I1204 18:16:32.645701 4733 generic.go:334] "Generic (PLEG): container finished" podID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerID="71fb1401d2180c86d22bd0a9a8f99d4c63a5b4c9c9edba91e55011192978304e" exitCode=0 Dec 04 18:16:32 crc kubenswrapper[4733]: I1204 18:16:32.645823 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7dh5" event={"ID":"e5009ea9-a807-4453-8f7b-73b987bc5aa7","Type":"ContainerDied","Data":"71fb1401d2180c86d22bd0a9a8f99d4c63a5b4c9c9edba91e55011192978304e"} Dec 04 18:16:34 crc kubenswrapper[4733]: I1204 18:16:34.667003 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7dh5" event={"ID":"e5009ea9-a807-4453-8f7b-73b987bc5aa7","Type":"ContainerStarted","Data":"4e59a86aca13ce1821c9d022a1290e6b504bcfb410037b75d0c428fea1e869ae"} Dec 04 18:16:34 crc kubenswrapper[4733]: I1204 18:16:34.695748 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f7dh5" podStartSLOduration=2.8659820160000002 podStartE2EDuration="5.695731086s" podCreationTimestamp="2025-12-04 18:16:29 +0000 UTC" firstStartedPulling="2025-12-04 18:16:30.623298822 +0000 UTC m=+2252.578659878" lastFinishedPulling="2025-12-04 18:16:33.453047862 +0000 UTC m=+2255.408408948" observedRunningTime="2025-12-04 18:16:34.695368156 +0000 UTC m=+2256.650729212" watchObservedRunningTime="2025-12-04 18:16:34.695731086 +0000 UTC m=+2256.651092142" Dec 04 18:16:39 crc kubenswrapper[4733]: I1204 18:16:39.914692 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:39 crc kubenswrapper[4733]: I1204 18:16:39.915302 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:39 crc kubenswrapper[4733]: I1204 18:16:39.999722 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:40 crc kubenswrapper[4733]: I1204 18:16:40.799069 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:40 crc kubenswrapper[4733]: I1204 18:16:40.862509 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f7dh5"] Dec 04 18:16:42 crc kubenswrapper[4733]: I1204 18:16:42.739210 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f7dh5" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerName="registry-server" containerID="cri-o://4e59a86aca13ce1821c9d022a1290e6b504bcfb410037b75d0c428fea1e869ae" gracePeriod=2 Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.750262 4733 generic.go:334] "Generic (PLEG): container finished" podID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerID="4e59a86aca13ce1821c9d022a1290e6b504bcfb410037b75d0c428fea1e869ae" exitCode=0 Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.750324 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7dh5" event={"ID":"e5009ea9-a807-4453-8f7b-73b987bc5aa7","Type":"ContainerDied","Data":"4e59a86aca13ce1821c9d022a1290e6b504bcfb410037b75d0c428fea1e869ae"} Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.750578 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f7dh5" event={"ID":"e5009ea9-a807-4453-8f7b-73b987bc5aa7","Type":"ContainerDied","Data":"5ad22598bb8428fac98b2f9251c7eb9699b640307cab9535519929517fb9470a"} Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.750600 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ad22598bb8428fac98b2f9251c7eb9699b640307cab9535519929517fb9470a" Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.800989 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.864476 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-catalog-content\") pod \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.864648 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2zdn\" (UniqueName: \"kubernetes.io/projected/e5009ea9-a807-4453-8f7b-73b987bc5aa7-kube-api-access-v2zdn\") pod \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.864780 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-utilities\") pod \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\" (UID: \"e5009ea9-a807-4453-8f7b-73b987bc5aa7\") " Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.866226 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-utilities" (OuterVolumeSpecName: "utilities") pod "e5009ea9-a807-4453-8f7b-73b987bc5aa7" (UID: "e5009ea9-a807-4453-8f7b-73b987bc5aa7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.872286 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5009ea9-a807-4453-8f7b-73b987bc5aa7-kube-api-access-v2zdn" (OuterVolumeSpecName: "kube-api-access-v2zdn") pod "e5009ea9-a807-4453-8f7b-73b987bc5aa7" (UID: "e5009ea9-a807-4453-8f7b-73b987bc5aa7"). InnerVolumeSpecName "kube-api-access-v2zdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.946966 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5009ea9-a807-4453-8f7b-73b987bc5aa7" (UID: "e5009ea9-a807-4453-8f7b-73b987bc5aa7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.966644 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2zdn\" (UniqueName: \"kubernetes.io/projected/e5009ea9-a807-4453-8f7b-73b987bc5aa7-kube-api-access-v2zdn\") on node \"crc\" DevicePath \"\"" Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.966691 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:16:43 crc kubenswrapper[4733]: I1204 18:16:43.966740 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5009ea9-a807-4453-8f7b-73b987bc5aa7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:16:44 crc kubenswrapper[4733]: I1204 18:16:44.760510 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f7dh5" Dec 04 18:16:44 crc kubenswrapper[4733]: I1204 18:16:44.793864 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f7dh5"] Dec 04 18:16:44 crc kubenswrapper[4733]: I1204 18:16:44.807145 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f7dh5"] Dec 04 18:16:45 crc kubenswrapper[4733]: I1204 18:16:45.362105 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:16:45 crc kubenswrapper[4733]: I1204 18:16:45.362220 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:16:46 crc kubenswrapper[4733]: I1204 18:16:46.349487 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" path="/var/lib/kubelet/pods/e5009ea9-a807-4453-8f7b-73b987bc5aa7/volumes" Dec 04 18:17:15 crc kubenswrapper[4733]: I1204 18:17:15.361680 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:17:15 crc kubenswrapper[4733]: I1204 18:17:15.362571 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:17:45 crc kubenswrapper[4733]: I1204 18:17:45.361911 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:17:45 crc kubenswrapper[4733]: I1204 18:17:45.362703 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:17:45 crc kubenswrapper[4733]: I1204 18:17:45.362772 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:17:45 crc kubenswrapper[4733]: I1204 18:17:45.364269 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:17:45 crc kubenswrapper[4733]: I1204 18:17:45.364457 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" gracePeriod=600 Dec 04 18:17:45 crc kubenswrapper[4733]: E1204 18:17:45.500776 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:17:46 crc kubenswrapper[4733]: I1204 18:17:46.380618 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" exitCode=0 Dec 04 18:17:46 crc kubenswrapper[4733]: I1204 18:17:46.380670 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c"} Dec 04 18:17:46 crc kubenswrapper[4733]: I1204 18:17:46.381466 4733 scope.go:117] "RemoveContainer" containerID="247e7e47e31412293bb049df2983c101d961b202a8b6d949ad0f187702a52ddf" Dec 04 18:17:46 crc kubenswrapper[4733]: I1204 18:17:46.382276 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:17:46 crc kubenswrapper[4733]: E1204 18:17:46.382775 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:17:57 crc kubenswrapper[4733]: I1204 18:17:57.336476 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:17:57 crc kubenswrapper[4733]: E1204 18:17:57.337620 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:18:12 crc kubenswrapper[4733]: I1204 18:18:12.336288 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:18:12 crc kubenswrapper[4733]: E1204 18:18:12.337375 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:18:25 crc kubenswrapper[4733]: I1204 18:18:25.336161 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:18:25 crc kubenswrapper[4733]: E1204 18:18:25.337199 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:18:40 crc kubenswrapper[4733]: I1204 18:18:40.336438 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:18:40 crc kubenswrapper[4733]: E1204 18:18:40.337927 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:18:53 crc kubenswrapper[4733]: I1204 18:18:53.336552 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:18:53 crc kubenswrapper[4733]: E1204 18:18:53.337919 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:19:05 crc kubenswrapper[4733]: I1204 18:19:05.336210 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:19:05 crc kubenswrapper[4733]: E1204 18:19:05.337409 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:19:17 crc kubenswrapper[4733]: I1204 18:19:17.335543 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:19:17 crc kubenswrapper[4733]: E1204 18:19:17.336330 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:19:19 crc kubenswrapper[4733]: I1204 18:19:19.238374 4733 scope.go:117] "RemoveContainer" containerID="b1589c56e9b6e9d06d75d9946fb2883989bcadad43431e8e27754b10433f4ab4" Dec 04 18:19:19 crc kubenswrapper[4733]: I1204 18:19:19.266873 4733 scope.go:117] "RemoveContainer" containerID="5bdca859ded93a4b4302771eefd90c01607de3634c4023a179b2c453a9fa862b" Dec 04 18:19:19 crc kubenswrapper[4733]: I1204 18:19:19.310297 4733 scope.go:117] "RemoveContainer" containerID="f6838dae9725c94d4ba9205ac14da84db281dca7690ff8d1d248c08c30d91ec9" Dec 04 18:19:28 crc kubenswrapper[4733]: I1204 18:19:28.341991 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:19:28 crc kubenswrapper[4733]: E1204 18:19:28.343225 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:19:39 crc kubenswrapper[4733]: I1204 18:19:39.336128 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:19:39 crc kubenswrapper[4733]: E1204 18:19:39.337197 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:19:52 crc kubenswrapper[4733]: I1204 18:19:52.335536 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:19:52 crc kubenswrapper[4733]: E1204 18:19:52.336562 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:20:04 crc kubenswrapper[4733]: I1204 18:20:04.336700 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:20:04 crc kubenswrapper[4733]: E1204 18:20:04.338376 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:20:19 crc kubenswrapper[4733]: I1204 18:20:19.373208 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:20:19 crc kubenswrapper[4733]: E1204 18:20:19.374246 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.440591 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f2qtr"] Dec 04 18:20:25 crc kubenswrapper[4733]: E1204 18:20:25.441615 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerName="extract-content" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.441636 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerName="extract-content" Dec 04 18:20:25 crc kubenswrapper[4733]: E1204 18:20:25.441665 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerName="extract-utilities" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.441679 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerName="extract-utilities" Dec 04 18:20:25 crc kubenswrapper[4733]: E1204 18:20:25.441718 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerName="registry-server" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.441732 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerName="registry-server" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.442023 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5009ea9-a807-4453-8f7b-73b987bc5aa7" containerName="registry-server" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.443861 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.466236 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2qtr"] Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.563292 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-catalog-content\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.563429 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-utilities\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.563501 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rclsx\" (UniqueName: \"kubernetes.io/projected/33c7180e-3511-41ef-93d2-bb3bb9073864-kube-api-access-rclsx\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.664377 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-catalog-content\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.664509 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-utilities\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.664942 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-catalog-content\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.665006 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-utilities\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.665145 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rclsx\" (UniqueName: \"kubernetes.io/projected/33c7180e-3511-41ef-93d2-bb3bb9073864-kube-api-access-rclsx\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.688382 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rclsx\" (UniqueName: \"kubernetes.io/projected/33c7180e-3511-41ef-93d2-bb3bb9073864-kube-api-access-rclsx\") pod \"redhat-marketplace-f2qtr\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:25 crc kubenswrapper[4733]: I1204 18:20:25.802744 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:26 crc kubenswrapper[4733]: I1204 18:20:26.305322 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2qtr"] Dec 04 18:20:26 crc kubenswrapper[4733]: I1204 18:20:26.953336 4733 generic.go:334] "Generic (PLEG): container finished" podID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerID="e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2" exitCode=0 Dec 04 18:20:26 crc kubenswrapper[4733]: I1204 18:20:26.953751 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2qtr" event={"ID":"33c7180e-3511-41ef-93d2-bb3bb9073864","Type":"ContainerDied","Data":"e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2"} Dec 04 18:20:26 crc kubenswrapper[4733]: I1204 18:20:26.953822 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2qtr" event={"ID":"33c7180e-3511-41ef-93d2-bb3bb9073864","Type":"ContainerStarted","Data":"79fabea78b00f94a653a011836f25f78a8faf47b804b64b7320091721126ccdc"} Dec 04 18:20:27 crc kubenswrapper[4733]: I1204 18:20:27.964608 4733 generic.go:334] "Generic (PLEG): container finished" podID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerID="d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09" exitCode=0 Dec 04 18:20:27 crc kubenswrapper[4733]: I1204 18:20:27.964745 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2qtr" event={"ID":"33c7180e-3511-41ef-93d2-bb3bb9073864","Type":"ContainerDied","Data":"d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09"} Dec 04 18:20:28 crc kubenswrapper[4733]: I1204 18:20:28.974297 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2qtr" event={"ID":"33c7180e-3511-41ef-93d2-bb3bb9073864","Type":"ContainerStarted","Data":"34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f"} Dec 04 18:20:28 crc kubenswrapper[4733]: I1204 18:20:28.995250 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f2qtr" podStartSLOduration=2.46687901 podStartE2EDuration="3.995235006s" podCreationTimestamp="2025-12-04 18:20:25 +0000 UTC" firstStartedPulling="2025-12-04 18:20:26.956146957 +0000 UTC m=+2488.911508043" lastFinishedPulling="2025-12-04 18:20:28.484502983 +0000 UTC m=+2490.439864039" observedRunningTime="2025-12-04 18:20:28.990426337 +0000 UTC m=+2490.945787423" watchObservedRunningTime="2025-12-04 18:20:28.995235006 +0000 UTC m=+2490.950596052" Dec 04 18:20:30 crc kubenswrapper[4733]: I1204 18:20:30.336093 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:20:30 crc kubenswrapper[4733]: E1204 18:20:30.336876 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:20:35 crc kubenswrapper[4733]: I1204 18:20:35.803880 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:35 crc kubenswrapper[4733]: I1204 18:20:35.805597 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:35 crc kubenswrapper[4733]: I1204 18:20:35.865012 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:36 crc kubenswrapper[4733]: I1204 18:20:36.095633 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:37 crc kubenswrapper[4733]: I1204 18:20:37.194606 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2qtr"] Dec 04 18:20:38 crc kubenswrapper[4733]: I1204 18:20:38.041477 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f2qtr" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerName="registry-server" containerID="cri-o://34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f" gracePeriod=2 Dec 04 18:20:38 crc kubenswrapper[4733]: I1204 18:20:38.993175 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.049438 4733 generic.go:334] "Generic (PLEG): container finished" podID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerID="34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f" exitCode=0 Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.049477 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2qtr" event={"ID":"33c7180e-3511-41ef-93d2-bb3bb9073864","Type":"ContainerDied","Data":"34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f"} Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.049486 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2qtr" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.049502 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2qtr" event={"ID":"33c7180e-3511-41ef-93d2-bb3bb9073864","Type":"ContainerDied","Data":"79fabea78b00f94a653a011836f25f78a8faf47b804b64b7320091721126ccdc"} Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.049517 4733 scope.go:117] "RemoveContainer" containerID="34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.070991 4733 scope.go:117] "RemoveContainer" containerID="d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.075067 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-catalog-content\") pod \"33c7180e-3511-41ef-93d2-bb3bb9073864\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.075133 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rclsx\" (UniqueName: \"kubernetes.io/projected/33c7180e-3511-41ef-93d2-bb3bb9073864-kube-api-access-rclsx\") pod \"33c7180e-3511-41ef-93d2-bb3bb9073864\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.075156 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-utilities\") pod \"33c7180e-3511-41ef-93d2-bb3bb9073864\" (UID: \"33c7180e-3511-41ef-93d2-bb3bb9073864\") " Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.076335 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-utilities" (OuterVolumeSpecName: "utilities") pod "33c7180e-3511-41ef-93d2-bb3bb9073864" (UID: "33c7180e-3511-41ef-93d2-bb3bb9073864"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.094056 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33c7180e-3511-41ef-93d2-bb3bb9073864-kube-api-access-rclsx" (OuterVolumeSpecName: "kube-api-access-rclsx") pod "33c7180e-3511-41ef-93d2-bb3bb9073864" (UID: "33c7180e-3511-41ef-93d2-bb3bb9073864"). InnerVolumeSpecName "kube-api-access-rclsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.100062 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33c7180e-3511-41ef-93d2-bb3bb9073864" (UID: "33c7180e-3511-41ef-93d2-bb3bb9073864"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.112636 4733 scope.go:117] "RemoveContainer" containerID="e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.137648 4733 scope.go:117] "RemoveContainer" containerID="34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f" Dec 04 18:20:39 crc kubenswrapper[4733]: E1204 18:20:39.137987 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f\": container with ID starting with 34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f not found: ID does not exist" containerID="34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.138012 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f"} err="failed to get container status \"34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f\": rpc error: code = NotFound desc = could not find container \"34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f\": container with ID starting with 34e242dcd9e21416653205e976dab59dcd5f3d0f33e26b59025e55f46b09df7f not found: ID does not exist" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.138030 4733 scope.go:117] "RemoveContainer" containerID="d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09" Dec 04 18:20:39 crc kubenswrapper[4733]: E1204 18:20:39.138237 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09\": container with ID starting with d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09 not found: ID does not exist" containerID="d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.138251 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09"} err="failed to get container status \"d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09\": rpc error: code = NotFound desc = could not find container \"d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09\": container with ID starting with d258ec6c7a2fa32510af327a81686f206b5944d69201d989a43ac54479c1bf09 not found: ID does not exist" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.138262 4733 scope.go:117] "RemoveContainer" containerID="e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2" Dec 04 18:20:39 crc kubenswrapper[4733]: E1204 18:20:39.138446 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2\": container with ID starting with e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2 not found: ID does not exist" containerID="e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.138461 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2"} err="failed to get container status \"e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2\": rpc error: code = NotFound desc = could not find container \"e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2\": container with ID starting with e50bcf21acbbbfb598d689fa85c2504c1256e9d91bca5a7f9643f9d2325e99c2 not found: ID does not exist" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.177705 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.177746 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rclsx\" (UniqueName: \"kubernetes.io/projected/33c7180e-3511-41ef-93d2-bb3bb9073864-kube-api-access-rclsx\") on node \"crc\" DevicePath \"\"" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.177761 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c7180e-3511-41ef-93d2-bb3bb9073864-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.391981 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2qtr"] Dec 04 18:20:39 crc kubenswrapper[4733]: I1204 18:20:39.396922 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2qtr"] Dec 04 18:20:40 crc kubenswrapper[4733]: I1204 18:20:40.350861 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" path="/var/lib/kubelet/pods/33c7180e-3511-41ef-93d2-bb3bb9073864/volumes" Dec 04 18:20:44 crc kubenswrapper[4733]: I1204 18:20:44.335829 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:20:44 crc kubenswrapper[4733]: E1204 18:20:44.336389 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:20:58 crc kubenswrapper[4733]: I1204 18:20:58.345400 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:20:58 crc kubenswrapper[4733]: E1204 18:20:58.346615 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:21:13 crc kubenswrapper[4733]: I1204 18:21:13.335568 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:21:13 crc kubenswrapper[4733]: E1204 18:21:13.336728 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:21:27 crc kubenswrapper[4733]: I1204 18:21:27.335699 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:21:27 crc kubenswrapper[4733]: E1204 18:21:27.337109 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:21:38 crc kubenswrapper[4733]: I1204 18:21:38.340408 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:21:38 crc kubenswrapper[4733]: E1204 18:21:38.341124 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:21:50 crc kubenswrapper[4733]: I1204 18:21:50.335779 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:21:50 crc kubenswrapper[4733]: E1204 18:21:50.337196 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:22:01 crc kubenswrapper[4733]: I1204 18:22:01.335635 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:22:01 crc kubenswrapper[4733]: E1204 18:22:01.336750 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:22:15 crc kubenswrapper[4733]: I1204 18:22:15.335196 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:22:15 crc kubenswrapper[4733]: E1204 18:22:15.336231 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:22:28 crc kubenswrapper[4733]: I1204 18:22:28.345391 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:22:28 crc kubenswrapper[4733]: E1204 18:22:28.346665 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:22:41 crc kubenswrapper[4733]: I1204 18:22:41.335987 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:22:41 crc kubenswrapper[4733]: E1204 18:22:41.337399 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:22:55 crc kubenswrapper[4733]: I1204 18:22:55.335991 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:22:56 crc kubenswrapper[4733]: I1204 18:22:56.377325 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"b5f1ecba5e7b2837bbaec97ab430bfa258d3ad2ba853e372e2f7f1a6712c5d59"} Dec 04 18:23:19 crc kubenswrapper[4733]: I1204 18:23:19.447413 4733 scope.go:117] "RemoveContainer" containerID="71fb1401d2180c86d22bd0a9a8f99d4c63a5b4c9c9edba91e55011192978304e" Dec 04 18:23:19 crc kubenswrapper[4733]: I1204 18:23:19.481143 4733 scope.go:117] "RemoveContainer" containerID="5fb225c1eeb12ed62644709887ebe8f8dd4d6600d732030c2ef37fe40df1aaa8" Dec 04 18:23:19 crc kubenswrapper[4733]: I1204 18:23:19.530955 4733 scope.go:117] "RemoveContainer" containerID="4e59a86aca13ce1821c9d022a1290e6b504bcfb410037b75d0c428fea1e869ae" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.475279 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vr44f"] Dec 04 18:23:42 crc kubenswrapper[4733]: E1204 18:23:42.477725 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerName="registry-server" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.477829 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerName="registry-server" Dec 04 18:23:42 crc kubenswrapper[4733]: E1204 18:23:42.477885 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerName="extract-utilities" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.477901 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerName="extract-utilities" Dec 04 18:23:42 crc kubenswrapper[4733]: E1204 18:23:42.477920 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerName="extract-content" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.477932 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerName="extract-content" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.478207 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="33c7180e-3511-41ef-93d2-bb3bb9073864" containerName="registry-server" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.480403 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.491313 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vr44f"] Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.539494 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-catalog-content\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.539560 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-utilities\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.539639 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmgtt\" (UniqueName: \"kubernetes.io/projected/68f5f4a9-adb5-491e-8bf9-3c700d7df631-kube-api-access-kmgtt\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.641949 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmgtt\" (UniqueName: \"kubernetes.io/projected/68f5f4a9-adb5-491e-8bf9-3c700d7df631-kube-api-access-kmgtt\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.642113 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-catalog-content\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.642177 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-utilities\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.642670 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-catalog-content\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.642753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-utilities\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.669845 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmgtt\" (UniqueName: \"kubernetes.io/projected/68f5f4a9-adb5-491e-8bf9-3c700d7df631-kube-api-access-kmgtt\") pod \"community-operators-vr44f\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:42 crc kubenswrapper[4733]: I1204 18:23:42.823517 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:43 crc kubenswrapper[4733]: I1204 18:23:43.419086 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vr44f"] Dec 04 18:23:43 crc kubenswrapper[4733]: I1204 18:23:43.814413 4733 generic.go:334] "Generic (PLEG): container finished" podID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerID="1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978" exitCode=0 Dec 04 18:23:43 crc kubenswrapper[4733]: I1204 18:23:43.814499 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr44f" event={"ID":"68f5f4a9-adb5-491e-8bf9-3c700d7df631","Type":"ContainerDied","Data":"1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978"} Dec 04 18:23:43 crc kubenswrapper[4733]: I1204 18:23:43.814878 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr44f" event={"ID":"68f5f4a9-adb5-491e-8bf9-3c700d7df631","Type":"ContainerStarted","Data":"ca572ebb18f12f1acce3a0dfb42be423741b0624efc319781b19e304573ae023"} Dec 04 18:23:43 crc kubenswrapper[4733]: I1204 18:23:43.818552 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:23:44 crc kubenswrapper[4733]: I1204 18:23:44.829131 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr44f" event={"ID":"68f5f4a9-adb5-491e-8bf9-3c700d7df631","Type":"ContainerStarted","Data":"5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5"} Dec 04 18:23:45 crc kubenswrapper[4733]: I1204 18:23:45.839996 4733 generic.go:334] "Generic (PLEG): container finished" podID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerID="5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5" exitCode=0 Dec 04 18:23:45 crc kubenswrapper[4733]: I1204 18:23:45.840085 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr44f" event={"ID":"68f5f4a9-adb5-491e-8bf9-3c700d7df631","Type":"ContainerDied","Data":"5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5"} Dec 04 18:23:46 crc kubenswrapper[4733]: I1204 18:23:46.850474 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr44f" event={"ID":"68f5f4a9-adb5-491e-8bf9-3c700d7df631","Type":"ContainerStarted","Data":"6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b"} Dec 04 18:23:52 crc kubenswrapper[4733]: I1204 18:23:52.823782 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:52 crc kubenswrapper[4733]: I1204 18:23:52.824099 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:52 crc kubenswrapper[4733]: I1204 18:23:52.899475 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:52 crc kubenswrapper[4733]: I1204 18:23:52.930546 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vr44f" podStartSLOduration=8.515144229 podStartE2EDuration="10.930464951s" podCreationTimestamp="2025-12-04 18:23:42 +0000 UTC" firstStartedPulling="2025-12-04 18:23:43.818051762 +0000 UTC m=+2685.773412848" lastFinishedPulling="2025-12-04 18:23:46.233372484 +0000 UTC m=+2688.188733570" observedRunningTime="2025-12-04 18:23:46.870087067 +0000 UTC m=+2688.825448153" watchObservedRunningTime="2025-12-04 18:23:52.930464951 +0000 UTC m=+2694.885826067" Dec 04 18:23:52 crc kubenswrapper[4733]: I1204 18:23:52.976078 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:53 crc kubenswrapper[4733]: I1204 18:23:53.149614 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vr44f"] Dec 04 18:23:54 crc kubenswrapper[4733]: I1204 18:23:54.924136 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vr44f" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerName="registry-server" containerID="cri-o://6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b" gracePeriod=2 Dec 04 18:23:55 crc kubenswrapper[4733]: I1204 18:23:55.864273 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:55 crc kubenswrapper[4733]: I1204 18:23:55.941172 4733 generic.go:334] "Generic (PLEG): container finished" podID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerID="6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b" exitCode=0 Dec 04 18:23:55 crc kubenswrapper[4733]: I1204 18:23:55.941233 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr44f" event={"ID":"68f5f4a9-adb5-491e-8bf9-3c700d7df631","Type":"ContainerDied","Data":"6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b"} Dec 04 18:23:55 crc kubenswrapper[4733]: I1204 18:23:55.941265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr44f" event={"ID":"68f5f4a9-adb5-491e-8bf9-3c700d7df631","Type":"ContainerDied","Data":"ca572ebb18f12f1acce3a0dfb42be423741b0624efc319781b19e304573ae023"} Dec 04 18:23:55 crc kubenswrapper[4733]: I1204 18:23:55.941305 4733 scope.go:117] "RemoveContainer" containerID="6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b" Dec 04 18:23:55 crc kubenswrapper[4733]: I1204 18:23:55.941504 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr44f" Dec 04 18:23:55 crc kubenswrapper[4733]: I1204 18:23:55.990787 4733 scope.go:117] "RemoveContainer" containerID="5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.006425 4733 scope.go:117] "RemoveContainer" containerID="1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.032728 4733 scope.go:117] "RemoveContainer" containerID="6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b" Dec 04 18:23:56 crc kubenswrapper[4733]: E1204 18:23:56.033204 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b\": container with ID starting with 6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b not found: ID does not exist" containerID="6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.033251 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b"} err="failed to get container status \"6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b\": rpc error: code = NotFound desc = could not find container \"6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b\": container with ID starting with 6a5c816c389d7318af3f2744176ec70420ad373457725429b6922830d50f275b not found: ID does not exist" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.033317 4733 scope.go:117] "RemoveContainer" containerID="5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5" Dec 04 18:23:56 crc kubenswrapper[4733]: E1204 18:23:56.033647 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5\": container with ID starting with 5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5 not found: ID does not exist" containerID="5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.033679 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5"} err="failed to get container status \"5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5\": rpc error: code = NotFound desc = could not find container \"5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5\": container with ID starting with 5ad4bd81a6764654a6363203b874dda91697488cfe6ebc943099cded956759c5 not found: ID does not exist" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.033700 4733 scope.go:117] "RemoveContainer" containerID="1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978" Dec 04 18:23:56 crc kubenswrapper[4733]: E1204 18:23:56.034019 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978\": container with ID starting with 1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978 not found: ID does not exist" containerID="1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.034081 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978"} err="failed to get container status \"1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978\": rpc error: code = NotFound desc = could not find container \"1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978\": container with ID starting with 1f26ac374ddcba1c692c624424ef30a77f148f0a93d14fde014efe85c55a5978 not found: ID does not exist" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.051046 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-utilities\") pod \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.051271 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmgtt\" (UniqueName: \"kubernetes.io/projected/68f5f4a9-adb5-491e-8bf9-3c700d7df631-kube-api-access-kmgtt\") pod \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.051408 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-catalog-content\") pod \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\" (UID: \"68f5f4a9-adb5-491e-8bf9-3c700d7df631\") " Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.052077 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-utilities" (OuterVolumeSpecName: "utilities") pod "68f5f4a9-adb5-491e-8bf9-3c700d7df631" (UID: "68f5f4a9-adb5-491e-8bf9-3c700d7df631"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.061204 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68f5f4a9-adb5-491e-8bf9-3c700d7df631-kube-api-access-kmgtt" (OuterVolumeSpecName: "kube-api-access-kmgtt") pod "68f5f4a9-adb5-491e-8bf9-3c700d7df631" (UID: "68f5f4a9-adb5-491e-8bf9-3c700d7df631"). InnerVolumeSpecName "kube-api-access-kmgtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.120585 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68f5f4a9-adb5-491e-8bf9-3c700d7df631" (UID: "68f5f4a9-adb5-491e-8bf9-3c700d7df631"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.153313 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.153509 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmgtt\" (UniqueName: \"kubernetes.io/projected/68f5f4a9-adb5-491e-8bf9-3c700d7df631-kube-api-access-kmgtt\") on node \"crc\" DevicePath \"\"" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.153536 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f5f4a9-adb5-491e-8bf9-3c700d7df631-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.292819 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vr44f"] Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.303689 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vr44f"] Dec 04 18:23:56 crc kubenswrapper[4733]: I1204 18:23:56.354318 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" path="/var/lib/kubelet/pods/68f5f4a9-adb5-491e-8bf9-3c700d7df631/volumes" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.531209 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xz6rc"] Dec 04 18:24:05 crc kubenswrapper[4733]: E1204 18:24:05.532180 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerName="extract-content" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.532195 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerName="extract-content" Dec 04 18:24:05 crc kubenswrapper[4733]: E1204 18:24:05.532254 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerName="extract-utilities" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.532286 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerName="extract-utilities" Dec 04 18:24:05 crc kubenswrapper[4733]: E1204 18:24:05.532303 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerName="registry-server" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.532312 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerName="registry-server" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.532945 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f5f4a9-adb5-491e-8bf9-3c700d7df631" containerName="registry-server" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.534569 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.555418 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xz6rc"] Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.714501 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-catalog-content\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.714623 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-utilities\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.714671 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xmkh\" (UniqueName: \"kubernetes.io/projected/6504de9f-62e3-4bc6-978e-cb24bacbc326-kube-api-access-8xmkh\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.816410 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-catalog-content\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.816493 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-utilities\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.816515 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xmkh\" (UniqueName: \"kubernetes.io/projected/6504de9f-62e3-4bc6-978e-cb24bacbc326-kube-api-access-8xmkh\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.817039 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-utilities\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.817076 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-catalog-content\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.836475 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xmkh\" (UniqueName: \"kubernetes.io/projected/6504de9f-62e3-4bc6-978e-cb24bacbc326-kube-api-access-8xmkh\") pod \"redhat-operators-xz6rc\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:05 crc kubenswrapper[4733]: I1204 18:24:05.871052 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:06 crc kubenswrapper[4733]: I1204 18:24:06.124197 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xz6rc"] Dec 04 18:24:07 crc kubenswrapper[4733]: I1204 18:24:07.041380 4733 generic.go:334] "Generic (PLEG): container finished" podID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerID="3e8b05e421f1165d9bcaea5a2a9e4c1ed40429514e96556ef3c20876aadd91bf" exitCode=0 Dec 04 18:24:07 crc kubenswrapper[4733]: I1204 18:24:07.041615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xz6rc" event={"ID":"6504de9f-62e3-4bc6-978e-cb24bacbc326","Type":"ContainerDied","Data":"3e8b05e421f1165d9bcaea5a2a9e4c1ed40429514e96556ef3c20876aadd91bf"} Dec 04 18:24:07 crc kubenswrapper[4733]: I1204 18:24:07.041944 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xz6rc" event={"ID":"6504de9f-62e3-4bc6-978e-cb24bacbc326","Type":"ContainerStarted","Data":"7847c90e79b61467042877987cfe79c2a4754fca85721cb132bcdfe2e61bf86f"} Dec 04 18:24:08 crc kubenswrapper[4733]: I1204 18:24:08.053377 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xz6rc" event={"ID":"6504de9f-62e3-4bc6-978e-cb24bacbc326","Type":"ContainerStarted","Data":"41549b72d1448e0e91d1f606f712e598a0aff0af32126c9cbdf83b9681d0b7dc"} Dec 04 18:24:09 crc kubenswrapper[4733]: I1204 18:24:09.065434 4733 generic.go:334] "Generic (PLEG): container finished" podID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerID="41549b72d1448e0e91d1f606f712e598a0aff0af32126c9cbdf83b9681d0b7dc" exitCode=0 Dec 04 18:24:09 crc kubenswrapper[4733]: I1204 18:24:09.065494 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xz6rc" event={"ID":"6504de9f-62e3-4bc6-978e-cb24bacbc326","Type":"ContainerDied","Data":"41549b72d1448e0e91d1f606f712e598a0aff0af32126c9cbdf83b9681d0b7dc"} Dec 04 18:24:10 crc kubenswrapper[4733]: I1204 18:24:10.075617 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xz6rc" event={"ID":"6504de9f-62e3-4bc6-978e-cb24bacbc326","Type":"ContainerStarted","Data":"f200f13aaae9df61515a45f9b396d19cc018b75e2a78dd3919a693f6a4898fc3"} Dec 04 18:24:10 crc kubenswrapper[4733]: I1204 18:24:10.107727 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xz6rc" podStartSLOduration=2.671322846 podStartE2EDuration="5.107700399s" podCreationTimestamp="2025-12-04 18:24:05 +0000 UTC" firstStartedPulling="2025-12-04 18:24:07.043132014 +0000 UTC m=+2708.998493070" lastFinishedPulling="2025-12-04 18:24:09.479509567 +0000 UTC m=+2711.434870623" observedRunningTime="2025-12-04 18:24:10.100417441 +0000 UTC m=+2712.055778497" watchObservedRunningTime="2025-12-04 18:24:10.107700399 +0000 UTC m=+2712.063061485" Dec 04 18:24:15 crc kubenswrapper[4733]: I1204 18:24:15.871335 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:15 crc kubenswrapper[4733]: I1204 18:24:15.871855 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:16 crc kubenswrapper[4733]: I1204 18:24:16.923245 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xz6rc" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="registry-server" probeResult="failure" output=< Dec 04 18:24:16 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 18:24:16 crc kubenswrapper[4733]: > Dec 04 18:24:25 crc kubenswrapper[4733]: I1204 18:24:25.938130 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:26 crc kubenswrapper[4733]: I1204 18:24:26.034889 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:26 crc kubenswrapper[4733]: I1204 18:24:26.181053 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xz6rc"] Dec 04 18:24:27 crc kubenswrapper[4733]: I1204 18:24:27.225379 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xz6rc" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="registry-server" containerID="cri-o://f200f13aaae9df61515a45f9b396d19cc018b75e2a78dd3919a693f6a4898fc3" gracePeriod=2 Dec 04 18:24:28 crc kubenswrapper[4733]: I1204 18:24:28.236989 4733 generic.go:334] "Generic (PLEG): container finished" podID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerID="f200f13aaae9df61515a45f9b396d19cc018b75e2a78dd3919a693f6a4898fc3" exitCode=0 Dec 04 18:24:28 crc kubenswrapper[4733]: I1204 18:24:28.237048 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xz6rc" event={"ID":"6504de9f-62e3-4bc6-978e-cb24bacbc326","Type":"ContainerDied","Data":"f200f13aaae9df61515a45f9b396d19cc018b75e2a78dd3919a693f6a4898fc3"} Dec 04 18:24:28 crc kubenswrapper[4733]: I1204 18:24:28.849785 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:28 crc kubenswrapper[4733]: I1204 18:24:28.982788 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-catalog-content\") pod \"6504de9f-62e3-4bc6-978e-cb24bacbc326\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " Dec 04 18:24:28 crc kubenswrapper[4733]: I1204 18:24:28.982991 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xmkh\" (UniqueName: \"kubernetes.io/projected/6504de9f-62e3-4bc6-978e-cb24bacbc326-kube-api-access-8xmkh\") pod \"6504de9f-62e3-4bc6-978e-cb24bacbc326\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " Dec 04 18:24:28 crc kubenswrapper[4733]: I1204 18:24:28.983077 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-utilities\") pod \"6504de9f-62e3-4bc6-978e-cb24bacbc326\" (UID: \"6504de9f-62e3-4bc6-978e-cb24bacbc326\") " Dec 04 18:24:28 crc kubenswrapper[4733]: I1204 18:24:28.984266 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-utilities" (OuterVolumeSpecName: "utilities") pod "6504de9f-62e3-4bc6-978e-cb24bacbc326" (UID: "6504de9f-62e3-4bc6-978e-cb24bacbc326"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:24:28 crc kubenswrapper[4733]: I1204 18:24:28.991944 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6504de9f-62e3-4bc6-978e-cb24bacbc326-kube-api-access-8xmkh" (OuterVolumeSpecName: "kube-api-access-8xmkh") pod "6504de9f-62e3-4bc6-978e-cb24bacbc326" (UID: "6504de9f-62e3-4bc6-978e-cb24bacbc326"). InnerVolumeSpecName "kube-api-access-8xmkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.084785 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xmkh\" (UniqueName: \"kubernetes.io/projected/6504de9f-62e3-4bc6-978e-cb24bacbc326-kube-api-access-8xmkh\") on node \"crc\" DevicePath \"\"" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.084856 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.142691 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6504de9f-62e3-4bc6-978e-cb24bacbc326" (UID: "6504de9f-62e3-4bc6-978e-cb24bacbc326"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.185705 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6504de9f-62e3-4bc6-978e-cb24bacbc326-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.248439 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xz6rc" event={"ID":"6504de9f-62e3-4bc6-978e-cb24bacbc326","Type":"ContainerDied","Data":"7847c90e79b61467042877987cfe79c2a4754fca85721cb132bcdfe2e61bf86f"} Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.248535 4733 scope.go:117] "RemoveContainer" containerID="f200f13aaae9df61515a45f9b396d19cc018b75e2a78dd3919a693f6a4898fc3" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.248571 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xz6rc" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.293349 4733 scope.go:117] "RemoveContainer" containerID="41549b72d1448e0e91d1f606f712e598a0aff0af32126c9cbdf83b9681d0b7dc" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.336833 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xz6rc"] Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.337919 4733 scope.go:117] "RemoveContainer" containerID="3e8b05e421f1165d9bcaea5a2a9e4c1ed40429514e96556ef3c20876aadd91bf" Dec 04 18:24:29 crc kubenswrapper[4733]: I1204 18:24:29.345081 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xz6rc"] Dec 04 18:24:30 crc kubenswrapper[4733]: I1204 18:24:30.351770 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" path="/var/lib/kubelet/pods/6504de9f-62e3-4bc6-978e-cb24bacbc326/volumes" Dec 04 18:25:15 crc kubenswrapper[4733]: I1204 18:25:15.362102 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:25:15 crc kubenswrapper[4733]: I1204 18:25:15.362716 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:25:45 crc kubenswrapper[4733]: I1204 18:25:45.362394 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:25:45 crc kubenswrapper[4733]: I1204 18:25:45.363255 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:26:15 crc kubenswrapper[4733]: I1204 18:26:15.362291 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:26:15 crc kubenswrapper[4733]: I1204 18:26:15.363268 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:26:15 crc kubenswrapper[4733]: I1204 18:26:15.363340 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:26:15 crc kubenswrapper[4733]: I1204 18:26:15.364244 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5f1ecba5e7b2837bbaec97ab430bfa258d3ad2ba853e372e2f7f1a6712c5d59"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:26:15 crc kubenswrapper[4733]: I1204 18:26:15.364325 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://b5f1ecba5e7b2837bbaec97ab430bfa258d3ad2ba853e372e2f7f1a6712c5d59" gracePeriod=600 Dec 04 18:26:16 crc kubenswrapper[4733]: I1204 18:26:16.125905 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="b5f1ecba5e7b2837bbaec97ab430bfa258d3ad2ba853e372e2f7f1a6712c5d59" exitCode=0 Dec 04 18:26:16 crc kubenswrapper[4733]: I1204 18:26:16.125918 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"b5f1ecba5e7b2837bbaec97ab430bfa258d3ad2ba853e372e2f7f1a6712c5d59"} Dec 04 18:26:16 crc kubenswrapper[4733]: I1204 18:26:16.127427 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759"} Dec 04 18:26:16 crc kubenswrapper[4733]: I1204 18:26:16.127476 4733 scope.go:117] "RemoveContainer" containerID="06d61fdc09d19d0f947c14912fa7d258b8472aaa3cdfa91ad77a78e88d8e449c" Dec 04 18:26:35 crc kubenswrapper[4733]: I1204 18:26:35.918258 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cdbgr"] Dec 04 18:26:35 crc kubenswrapper[4733]: E1204 18:26:35.919529 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="extract-utilities" Dec 04 18:26:35 crc kubenswrapper[4733]: I1204 18:26:35.919547 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="extract-utilities" Dec 04 18:26:35 crc kubenswrapper[4733]: E1204 18:26:35.919573 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="extract-content" Dec 04 18:26:35 crc kubenswrapper[4733]: I1204 18:26:35.919583 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="extract-content" Dec 04 18:26:35 crc kubenswrapper[4733]: E1204 18:26:35.919601 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="registry-server" Dec 04 18:26:35 crc kubenswrapper[4733]: I1204 18:26:35.919611 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="registry-server" Dec 04 18:26:35 crc kubenswrapper[4733]: I1204 18:26:35.919779 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6504de9f-62e3-4bc6-978e-cb24bacbc326" containerName="registry-server" Dec 04 18:26:35 crc kubenswrapper[4733]: I1204 18:26:35.921204 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:35 crc kubenswrapper[4733]: I1204 18:26:35.932985 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cdbgr"] Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.072013 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761f493e-8e27-4b0d-9a01-09b53b38e142-catalog-content\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.072072 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761f493e-8e27-4b0d-9a01-09b53b38e142-utilities\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.072306 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdzq7\" (UniqueName: \"kubernetes.io/projected/761f493e-8e27-4b0d-9a01-09b53b38e142-kube-api-access-zdzq7\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.173874 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761f493e-8e27-4b0d-9a01-09b53b38e142-catalog-content\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.174195 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761f493e-8e27-4b0d-9a01-09b53b38e142-utilities\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.174276 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdzq7\" (UniqueName: \"kubernetes.io/projected/761f493e-8e27-4b0d-9a01-09b53b38e142-kube-api-access-zdzq7\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.174458 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761f493e-8e27-4b0d-9a01-09b53b38e142-catalog-content\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.174504 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761f493e-8e27-4b0d-9a01-09b53b38e142-utilities\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.194519 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdzq7\" (UniqueName: \"kubernetes.io/projected/761f493e-8e27-4b0d-9a01-09b53b38e142-kube-api-access-zdzq7\") pod \"certified-operators-cdbgr\" (UID: \"761f493e-8e27-4b0d-9a01-09b53b38e142\") " pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.240458 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:36 crc kubenswrapper[4733]: I1204 18:26:36.709500 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cdbgr"] Dec 04 18:26:37 crc kubenswrapper[4733]: I1204 18:26:37.306417 4733 generic.go:334] "Generic (PLEG): container finished" podID="761f493e-8e27-4b0d-9a01-09b53b38e142" containerID="d713aab7311e9dcbb6cddb111d0b5d16b0b7f80344d60f0791ffe9a3af11dde8" exitCode=0 Dec 04 18:26:37 crc kubenswrapper[4733]: I1204 18:26:37.306490 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdbgr" event={"ID":"761f493e-8e27-4b0d-9a01-09b53b38e142","Type":"ContainerDied","Data":"d713aab7311e9dcbb6cddb111d0b5d16b0b7f80344d60f0791ffe9a3af11dde8"} Dec 04 18:26:37 crc kubenswrapper[4733]: I1204 18:26:37.306534 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdbgr" event={"ID":"761f493e-8e27-4b0d-9a01-09b53b38e142","Type":"ContainerStarted","Data":"dee79b8d04dec280d1d42683fa364024bc39bf8a9374f785a18a5296e585d32c"} Dec 04 18:26:41 crc kubenswrapper[4733]: I1204 18:26:41.338569 4733 generic.go:334] "Generic (PLEG): container finished" podID="761f493e-8e27-4b0d-9a01-09b53b38e142" containerID="31e9cd4054a95b8ac3016987c42f43ebf2a20e106f43fcfed96dd4468c2a9045" exitCode=0 Dec 04 18:26:41 crc kubenswrapper[4733]: I1204 18:26:41.338624 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdbgr" event={"ID":"761f493e-8e27-4b0d-9a01-09b53b38e142","Type":"ContainerDied","Data":"31e9cd4054a95b8ac3016987c42f43ebf2a20e106f43fcfed96dd4468c2a9045"} Dec 04 18:26:42 crc kubenswrapper[4733]: I1204 18:26:42.362202 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdbgr" event={"ID":"761f493e-8e27-4b0d-9a01-09b53b38e142","Type":"ContainerStarted","Data":"6ef6a1fff06bad7808adf5e04999209ef716bce339fecdef7133dd101530e3eb"} Dec 04 18:26:42 crc kubenswrapper[4733]: I1204 18:26:42.406670 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cdbgr" podStartSLOduration=2.950638745 podStartE2EDuration="7.406644907s" podCreationTimestamp="2025-12-04 18:26:35 +0000 UTC" firstStartedPulling="2025-12-04 18:26:37.308605076 +0000 UTC m=+2859.263966162" lastFinishedPulling="2025-12-04 18:26:41.764611278 +0000 UTC m=+2863.719972324" observedRunningTime="2025-12-04 18:26:42.400351246 +0000 UTC m=+2864.355712292" watchObservedRunningTime="2025-12-04 18:26:42.406644907 +0000 UTC m=+2864.362005953" Dec 04 18:26:46 crc kubenswrapper[4733]: I1204 18:26:46.242567 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:46 crc kubenswrapper[4733]: I1204 18:26:46.244496 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:46 crc kubenswrapper[4733]: I1204 18:26:46.320332 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:46 crc kubenswrapper[4733]: I1204 18:26:46.440732 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cdbgr" Dec 04 18:26:46 crc kubenswrapper[4733]: I1204 18:26:46.515539 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cdbgr"] Dec 04 18:26:46 crc kubenswrapper[4733]: I1204 18:26:46.572260 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vgmxv"] Dec 04 18:26:46 crc kubenswrapper[4733]: I1204 18:26:46.572622 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vgmxv" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="registry-server" containerID="cri-o://f3f567439952f0036242a59996a3223171a1ae3541639697fb8c5a0b8a5a6cab" gracePeriod=2 Dec 04 18:26:47 crc kubenswrapper[4733]: I1204 18:26:47.399470 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerID="f3f567439952f0036242a59996a3223171a1ae3541639697fb8c5a0b8a5a6cab" exitCode=0 Dec 04 18:26:47 crc kubenswrapper[4733]: I1204 18:26:47.399515 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vgmxv" event={"ID":"9d14d9e2-606c-46c2-8326-5794c29d0c32","Type":"ContainerDied","Data":"f3f567439952f0036242a59996a3223171a1ae3541639697fb8c5a0b8a5a6cab"} Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.101915 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.158505 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-catalog-content\") pod \"9d14d9e2-606c-46c2-8326-5794c29d0c32\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.158779 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-utilities\") pod \"9d14d9e2-606c-46c2-8326-5794c29d0c32\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.158868 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmwgr\" (UniqueName: \"kubernetes.io/projected/9d14d9e2-606c-46c2-8326-5794c29d0c32-kube-api-access-vmwgr\") pod \"9d14d9e2-606c-46c2-8326-5794c29d0c32\" (UID: \"9d14d9e2-606c-46c2-8326-5794c29d0c32\") " Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.165192 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d14d9e2-606c-46c2-8326-5794c29d0c32-kube-api-access-vmwgr" (OuterVolumeSpecName: "kube-api-access-vmwgr") pod "9d14d9e2-606c-46c2-8326-5794c29d0c32" (UID: "9d14d9e2-606c-46c2-8326-5794c29d0c32"). InnerVolumeSpecName "kube-api-access-vmwgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.167525 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmwgr\" (UniqueName: \"kubernetes.io/projected/9d14d9e2-606c-46c2-8326-5794c29d0c32-kube-api-access-vmwgr\") on node \"crc\" DevicePath \"\"" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.167865 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-utilities" (OuterVolumeSpecName: "utilities") pod "9d14d9e2-606c-46c2-8326-5794c29d0c32" (UID: "9d14d9e2-606c-46c2-8326-5794c29d0c32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.218986 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d14d9e2-606c-46c2-8326-5794c29d0c32" (UID: "9d14d9e2-606c-46c2-8326-5794c29d0c32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.269265 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.269307 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d14d9e2-606c-46c2-8326-5794c29d0c32-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.407919 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vgmxv" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.407976 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vgmxv" event={"ID":"9d14d9e2-606c-46c2-8326-5794c29d0c32","Type":"ContainerDied","Data":"12801190410ee0d93f5c083c271302f3d0662a0e95c2dcb4253b4ad88844e24b"} Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.408018 4733 scope.go:117] "RemoveContainer" containerID="f3f567439952f0036242a59996a3223171a1ae3541639697fb8c5a0b8a5a6cab" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.424491 4733 scope.go:117] "RemoveContainer" containerID="c66678c24e05da5f46aa54210b20c5576a61314666b6c82fce300f09a42fc49f" Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.433180 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vgmxv"] Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.439366 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vgmxv"] Dec 04 18:26:48 crc kubenswrapper[4733]: I1204 18:26:48.449188 4733 scope.go:117] "RemoveContainer" containerID="fd22bfd2e9b52af5a5a65011560ce7c1fafa380fb89acfc999fdddeef9ebab34" Dec 04 18:26:50 crc kubenswrapper[4733]: I1204 18:26:50.351892 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" path="/var/lib/kubelet/pods/9d14d9e2-606c-46c2-8326-5794c29d0c32/volumes" Dec 04 18:28:15 crc kubenswrapper[4733]: I1204 18:28:15.361963 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:28:15 crc kubenswrapper[4733]: I1204 18:28:15.362836 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:28:45 crc kubenswrapper[4733]: I1204 18:28:45.362456 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:28:45 crc kubenswrapper[4733]: I1204 18:28:45.363297 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.361843 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.362468 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.362524 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.363411 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.363510 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" gracePeriod=600 Dec 04 18:29:15 crc kubenswrapper[4733]: E1204 18:29:15.506787 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.655669 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" exitCode=0 Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.655736 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759"} Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.655789 4733 scope.go:117] "RemoveContainer" containerID="b5f1ecba5e7b2837bbaec97ab430bfa258d3ad2ba853e372e2f7f1a6712c5d59" Dec 04 18:29:15 crc kubenswrapper[4733]: I1204 18:29:15.656541 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:29:15 crc kubenswrapper[4733]: E1204 18:29:15.656968 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:29:28 crc kubenswrapper[4733]: I1204 18:29:28.340311 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:29:28 crc kubenswrapper[4733]: E1204 18:29:28.341177 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:29:43 crc kubenswrapper[4733]: I1204 18:29:43.335384 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:29:43 crc kubenswrapper[4733]: E1204 18:29:43.336541 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:29:57 crc kubenswrapper[4733]: I1204 18:29:57.336279 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:29:57 crc kubenswrapper[4733]: E1204 18:29:57.337548 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.159808 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4"] Dec 04 18:30:00 crc kubenswrapper[4733]: E1204 18:30:00.160918 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="registry-server" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.160939 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="registry-server" Dec 04 18:30:00 crc kubenswrapper[4733]: E1204 18:30:00.160970 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="extract-content" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.160982 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="extract-content" Dec 04 18:30:00 crc kubenswrapper[4733]: E1204 18:30:00.161004 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="extract-utilities" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.161015 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="extract-utilities" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.161252 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d14d9e2-606c-46c2-8326-5794c29d0c32" containerName="registry-server" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.162067 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.164599 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.167097 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.170525 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4"] Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.234951 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9d57197-10b5-4af4-b43a-3626efd322c3-config-volume\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.235012 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfq7s\" (UniqueName: \"kubernetes.io/projected/a9d57197-10b5-4af4-b43a-3626efd322c3-kube-api-access-zfq7s\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.235155 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9d57197-10b5-4af4-b43a-3626efd322c3-secret-volume\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.336267 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9d57197-10b5-4af4-b43a-3626efd322c3-config-volume\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.336356 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfq7s\" (UniqueName: \"kubernetes.io/projected/a9d57197-10b5-4af4-b43a-3626efd322c3-kube-api-access-zfq7s\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.336623 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9d57197-10b5-4af4-b43a-3626efd322c3-secret-volume\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.338067 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9d57197-10b5-4af4-b43a-3626efd322c3-config-volume\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.344977 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9d57197-10b5-4af4-b43a-3626efd322c3-secret-volume\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.359859 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfq7s\" (UniqueName: \"kubernetes.io/projected/a9d57197-10b5-4af4-b43a-3626efd322c3-kube-api-access-zfq7s\") pod \"collect-profiles-29414550-r7lf4\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.484572 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:00 crc kubenswrapper[4733]: I1204 18:30:00.940856 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4"] Dec 04 18:30:01 crc kubenswrapper[4733]: I1204 18:30:01.039487 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" event={"ID":"a9d57197-10b5-4af4-b43a-3626efd322c3","Type":"ContainerStarted","Data":"9600b9bf2a28104bf28990ada6aa806d107d79594b8ae97b9f1fe8d4c780d837"} Dec 04 18:30:02 crc kubenswrapper[4733]: I1204 18:30:02.050726 4733 generic.go:334] "Generic (PLEG): container finished" podID="a9d57197-10b5-4af4-b43a-3626efd322c3" containerID="d2745a6816337974701f65b2cae3728835beb963bd60f93c9802c2c97dadcd64" exitCode=0 Dec 04 18:30:02 crc kubenswrapper[4733]: I1204 18:30:02.050825 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" event={"ID":"a9d57197-10b5-4af4-b43a-3626efd322c3","Type":"ContainerDied","Data":"d2745a6816337974701f65b2cae3728835beb963bd60f93c9802c2c97dadcd64"} Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.376210 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.379312 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9d57197-10b5-4af4-b43a-3626efd322c3-config-volume\") pod \"a9d57197-10b5-4af4-b43a-3626efd322c3\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.379350 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9d57197-10b5-4af4-b43a-3626efd322c3-secret-volume\") pod \"a9d57197-10b5-4af4-b43a-3626efd322c3\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.379385 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfq7s\" (UniqueName: \"kubernetes.io/projected/a9d57197-10b5-4af4-b43a-3626efd322c3-kube-api-access-zfq7s\") pod \"a9d57197-10b5-4af4-b43a-3626efd322c3\" (UID: \"a9d57197-10b5-4af4-b43a-3626efd322c3\") " Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.380394 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d57197-10b5-4af4-b43a-3626efd322c3-config-volume" (OuterVolumeSpecName: "config-volume") pod "a9d57197-10b5-4af4-b43a-3626efd322c3" (UID: "a9d57197-10b5-4af4-b43a-3626efd322c3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.386162 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d57197-10b5-4af4-b43a-3626efd322c3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a9d57197-10b5-4af4-b43a-3626efd322c3" (UID: "a9d57197-10b5-4af4-b43a-3626efd322c3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.386356 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d57197-10b5-4af4-b43a-3626efd322c3-kube-api-access-zfq7s" (OuterVolumeSpecName: "kube-api-access-zfq7s") pod "a9d57197-10b5-4af4-b43a-3626efd322c3" (UID: "a9d57197-10b5-4af4-b43a-3626efd322c3"). InnerVolumeSpecName "kube-api-access-zfq7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.480315 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9d57197-10b5-4af4-b43a-3626efd322c3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.480345 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9d57197-10b5-4af4-b43a-3626efd322c3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 18:30:03 crc kubenswrapper[4733]: I1204 18:30:03.480357 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfq7s\" (UniqueName: \"kubernetes.io/projected/a9d57197-10b5-4af4-b43a-3626efd322c3-kube-api-access-zfq7s\") on node \"crc\" DevicePath \"\"" Dec 04 18:30:04 crc kubenswrapper[4733]: I1204 18:30:04.068781 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" event={"ID":"a9d57197-10b5-4af4-b43a-3626efd322c3","Type":"ContainerDied","Data":"9600b9bf2a28104bf28990ada6aa806d107d79594b8ae97b9f1fe8d4c780d837"} Dec 04 18:30:04 crc kubenswrapper[4733]: I1204 18:30:04.068855 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9600b9bf2a28104bf28990ada6aa806d107d79594b8ae97b9f1fe8d4c780d837" Dec 04 18:30:04 crc kubenswrapper[4733]: I1204 18:30:04.068900 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4" Dec 04 18:30:04 crc kubenswrapper[4733]: I1204 18:30:04.457626 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g"] Dec 04 18:30:04 crc kubenswrapper[4733]: I1204 18:30:04.467156 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414505-47z5g"] Dec 04 18:30:06 crc kubenswrapper[4733]: I1204 18:30:06.345592 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="535164e0-79d0-4b37-8e01-c84932cf56d6" path="/var/lib/kubelet/pods/535164e0-79d0-4b37-8e01-c84932cf56d6/volumes" Dec 04 18:30:09 crc kubenswrapper[4733]: I1204 18:30:09.335661 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:30:09 crc kubenswrapper[4733]: E1204 18:30:09.336043 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:30:19 crc kubenswrapper[4733]: I1204 18:30:19.776012 4733 scope.go:117] "RemoveContainer" containerID="0ffa215e4056fa6b8382d834afca207dc6aa1683cb1451763ab5ea6daba146f8" Dec 04 18:30:23 crc kubenswrapper[4733]: I1204 18:30:23.335186 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:30:23 crc kubenswrapper[4733]: E1204 18:30:23.336141 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:30:38 crc kubenswrapper[4733]: I1204 18:30:38.339641 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:30:38 crc kubenswrapper[4733]: E1204 18:30:38.340374 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:30:42 crc kubenswrapper[4733]: I1204 18:30:42.879974 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d99bh"] Dec 04 18:30:42 crc kubenswrapper[4733]: E1204 18:30:42.880459 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d57197-10b5-4af4-b43a-3626efd322c3" containerName="collect-profiles" Dec 04 18:30:42 crc kubenswrapper[4733]: I1204 18:30:42.880472 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d57197-10b5-4af4-b43a-3626efd322c3" containerName="collect-profiles" Dec 04 18:30:42 crc kubenswrapper[4733]: I1204 18:30:42.880606 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d57197-10b5-4af4-b43a-3626efd322c3" containerName="collect-profiles" Dec 04 18:30:42 crc kubenswrapper[4733]: I1204 18:30:42.881565 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:42 crc kubenswrapper[4733]: I1204 18:30:42.904714 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99bh"] Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.078402 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-utilities\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.078593 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-catalog-content\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.078638 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29ldw\" (UniqueName: \"kubernetes.io/projected/1220cdd1-b7bd-48be-bbd7-415053bf91eb-kube-api-access-29ldw\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.179523 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-catalog-content\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.179588 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29ldw\" (UniqueName: \"kubernetes.io/projected/1220cdd1-b7bd-48be-bbd7-415053bf91eb-kube-api-access-29ldw\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.179676 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-utilities\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.180320 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-catalog-content\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.180329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-utilities\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.201871 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29ldw\" (UniqueName: \"kubernetes.io/projected/1220cdd1-b7bd-48be-bbd7-415053bf91eb-kube-api-access-29ldw\") pod \"redhat-marketplace-d99bh\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.202939 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:43 crc kubenswrapper[4733]: I1204 18:30:43.679874 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99bh"] Dec 04 18:30:44 crc kubenswrapper[4733]: I1204 18:30:44.403301 4733 generic.go:334] "Generic (PLEG): container finished" podID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerID="63af3cccb559df36bf61090eeb1c78211fdb7be44dae4489c1eb3abe5960108f" exitCode=0 Dec 04 18:30:44 crc kubenswrapper[4733]: I1204 18:30:44.403483 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99bh" event={"ID":"1220cdd1-b7bd-48be-bbd7-415053bf91eb","Type":"ContainerDied","Data":"63af3cccb559df36bf61090eeb1c78211fdb7be44dae4489c1eb3abe5960108f"} Dec 04 18:30:44 crc kubenswrapper[4733]: I1204 18:30:44.403623 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99bh" event={"ID":"1220cdd1-b7bd-48be-bbd7-415053bf91eb","Type":"ContainerStarted","Data":"4bd5f27131f988ffc20a284899221c1542102d8b43260681e8d14609d7f08bc8"} Dec 04 18:30:44 crc kubenswrapper[4733]: I1204 18:30:44.405117 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:30:45 crc kubenswrapper[4733]: I1204 18:30:45.415826 4733 generic.go:334] "Generic (PLEG): container finished" podID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerID="5f9f1bed48b306e6e1f436775bcb9e3323cdfc74c6d92367bd78197b81e4fe44" exitCode=0 Dec 04 18:30:45 crc kubenswrapper[4733]: I1204 18:30:45.415898 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99bh" event={"ID":"1220cdd1-b7bd-48be-bbd7-415053bf91eb","Type":"ContainerDied","Data":"5f9f1bed48b306e6e1f436775bcb9e3323cdfc74c6d92367bd78197b81e4fe44"} Dec 04 18:30:46 crc kubenswrapper[4733]: I1204 18:30:46.424266 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99bh" event={"ID":"1220cdd1-b7bd-48be-bbd7-415053bf91eb","Type":"ContainerStarted","Data":"530fc5a85a31a0ea36876c613288b22073bfebc1b0f82787f0bba7f8ea6faa08"} Dec 04 18:30:46 crc kubenswrapper[4733]: I1204 18:30:46.448625 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d99bh" podStartSLOduration=3.003808008 podStartE2EDuration="4.448597953s" podCreationTimestamp="2025-12-04 18:30:42 +0000 UTC" firstStartedPulling="2025-12-04 18:30:44.404869687 +0000 UTC m=+3106.360230733" lastFinishedPulling="2025-12-04 18:30:45.849659602 +0000 UTC m=+3107.805020678" observedRunningTime="2025-12-04 18:30:46.44261143 +0000 UTC m=+3108.397972516" watchObservedRunningTime="2025-12-04 18:30:46.448597953 +0000 UTC m=+3108.403959019" Dec 04 18:30:49 crc kubenswrapper[4733]: I1204 18:30:49.336429 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:30:49 crc kubenswrapper[4733]: E1204 18:30:49.337781 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:30:53 crc kubenswrapper[4733]: I1204 18:30:53.203938 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:53 crc kubenswrapper[4733]: I1204 18:30:53.204327 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:53 crc kubenswrapper[4733]: I1204 18:30:53.252916 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:53 crc kubenswrapper[4733]: I1204 18:30:53.518146 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:53 crc kubenswrapper[4733]: I1204 18:30:53.564996 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99bh"] Dec 04 18:30:55 crc kubenswrapper[4733]: I1204 18:30:55.494164 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d99bh" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerName="registry-server" containerID="cri-o://530fc5a85a31a0ea36876c613288b22073bfebc1b0f82787f0bba7f8ea6faa08" gracePeriod=2 Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.503601 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99bh" event={"ID":"1220cdd1-b7bd-48be-bbd7-415053bf91eb","Type":"ContainerDied","Data":"530fc5a85a31a0ea36876c613288b22073bfebc1b0f82787f0bba7f8ea6faa08"} Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.503531 4733 generic.go:334] "Generic (PLEG): container finished" podID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerID="530fc5a85a31a0ea36876c613288b22073bfebc1b0f82787f0bba7f8ea6faa08" exitCode=0 Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.585096 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.720974 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-catalog-content\") pod \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.721095 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-utilities\") pod \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.721129 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29ldw\" (UniqueName: \"kubernetes.io/projected/1220cdd1-b7bd-48be-bbd7-415053bf91eb-kube-api-access-29ldw\") pod \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\" (UID: \"1220cdd1-b7bd-48be-bbd7-415053bf91eb\") " Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.722607 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-utilities" (OuterVolumeSpecName: "utilities") pod "1220cdd1-b7bd-48be-bbd7-415053bf91eb" (UID: "1220cdd1-b7bd-48be-bbd7-415053bf91eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.727465 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1220cdd1-b7bd-48be-bbd7-415053bf91eb-kube-api-access-29ldw" (OuterVolumeSpecName: "kube-api-access-29ldw") pod "1220cdd1-b7bd-48be-bbd7-415053bf91eb" (UID: "1220cdd1-b7bd-48be-bbd7-415053bf91eb"). InnerVolumeSpecName "kube-api-access-29ldw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.740213 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1220cdd1-b7bd-48be-bbd7-415053bf91eb" (UID: "1220cdd1-b7bd-48be-bbd7-415053bf91eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.823180 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.823218 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1220cdd1-b7bd-48be-bbd7-415053bf91eb-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:30:56 crc kubenswrapper[4733]: I1204 18:30:56.823236 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29ldw\" (UniqueName: \"kubernetes.io/projected/1220cdd1-b7bd-48be-bbd7-415053bf91eb-kube-api-access-29ldw\") on node \"crc\" DevicePath \"\"" Dec 04 18:30:57 crc kubenswrapper[4733]: I1204 18:30:57.515020 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d99bh" event={"ID":"1220cdd1-b7bd-48be-bbd7-415053bf91eb","Type":"ContainerDied","Data":"4bd5f27131f988ffc20a284899221c1542102d8b43260681e8d14609d7f08bc8"} Dec 04 18:30:57 crc kubenswrapper[4733]: I1204 18:30:57.515407 4733 scope.go:117] "RemoveContainer" containerID="530fc5a85a31a0ea36876c613288b22073bfebc1b0f82787f0bba7f8ea6faa08" Dec 04 18:30:57 crc kubenswrapper[4733]: I1204 18:30:57.515111 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d99bh" Dec 04 18:30:57 crc kubenswrapper[4733]: I1204 18:30:57.552453 4733 scope.go:117] "RemoveContainer" containerID="5f9f1bed48b306e6e1f436775bcb9e3323cdfc74c6d92367bd78197b81e4fe44" Dec 04 18:30:57 crc kubenswrapper[4733]: I1204 18:30:57.589443 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99bh"] Dec 04 18:30:57 crc kubenswrapper[4733]: I1204 18:30:57.597031 4733 scope.go:117] "RemoveContainer" containerID="63af3cccb559df36bf61090eeb1c78211fdb7be44dae4489c1eb3abe5960108f" Dec 04 18:30:57 crc kubenswrapper[4733]: I1204 18:30:57.618230 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d99bh"] Dec 04 18:30:58 crc kubenswrapper[4733]: I1204 18:30:58.351374 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" path="/var/lib/kubelet/pods/1220cdd1-b7bd-48be-bbd7-415053bf91eb/volumes" Dec 04 18:31:00 crc kubenswrapper[4733]: I1204 18:31:00.336555 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:31:00 crc kubenswrapper[4733]: E1204 18:31:00.337552 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:31:15 crc kubenswrapper[4733]: I1204 18:31:15.335982 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:31:15 crc kubenswrapper[4733]: E1204 18:31:15.337175 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:31:30 crc kubenswrapper[4733]: I1204 18:31:30.336751 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:31:30 crc kubenswrapper[4733]: E1204 18:31:30.338307 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:31:44 crc kubenswrapper[4733]: I1204 18:31:44.336154 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:31:44 crc kubenswrapper[4733]: E1204 18:31:44.337013 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:31:57 crc kubenswrapper[4733]: I1204 18:31:57.335565 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:31:57 crc kubenswrapper[4733]: E1204 18:31:57.336888 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:32:09 crc kubenswrapper[4733]: I1204 18:32:09.335376 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:32:09 crc kubenswrapper[4733]: E1204 18:32:09.336384 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:32:22 crc kubenswrapper[4733]: I1204 18:32:22.335706 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:32:22 crc kubenswrapper[4733]: E1204 18:32:22.336626 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:32:37 crc kubenswrapper[4733]: I1204 18:32:37.336070 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:32:37 crc kubenswrapper[4733]: E1204 18:32:37.336947 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:32:52 crc kubenswrapper[4733]: I1204 18:32:52.336578 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:32:52 crc kubenswrapper[4733]: E1204 18:32:52.337834 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:33:05 crc kubenswrapper[4733]: I1204 18:33:05.335343 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:33:05 crc kubenswrapper[4733]: E1204 18:33:05.336446 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:33:19 crc kubenswrapper[4733]: I1204 18:33:19.336206 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:33:19 crc kubenswrapper[4733]: E1204 18:33:19.337560 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:33:33 crc kubenswrapper[4733]: I1204 18:33:33.335528 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:33:33 crc kubenswrapper[4733]: E1204 18:33:33.336949 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:33:46 crc kubenswrapper[4733]: I1204 18:33:46.336104 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:33:46 crc kubenswrapper[4733]: E1204 18:33:46.337337 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:34:01 crc kubenswrapper[4733]: I1204 18:34:01.335165 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:34:01 crc kubenswrapper[4733]: E1204 18:34:01.336218 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.809994 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w978q"] Dec 04 18:34:09 crc kubenswrapper[4733]: E1204 18:34:09.810888 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerName="extract-utilities" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.810903 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerName="extract-utilities" Dec 04 18:34:09 crc kubenswrapper[4733]: E1204 18:34:09.810916 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerName="registry-server" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.810924 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerName="registry-server" Dec 04 18:34:09 crc kubenswrapper[4733]: E1204 18:34:09.810940 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerName="extract-content" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.810948 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerName="extract-content" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.811172 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1220cdd1-b7bd-48be-bbd7-415053bf91eb" containerName="registry-server" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.812466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.835168 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w978q"] Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.957190 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m7fg\" (UniqueName: \"kubernetes.io/projected/093c64a9-5cfc-4307-bc50-b8a50c263b48-kube-api-access-2m7fg\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.957282 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-utilities\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:09 crc kubenswrapper[4733]: I1204 18:34:09.957329 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-catalog-content\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:10 crc kubenswrapper[4733]: I1204 18:34:10.058685 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m7fg\" (UniqueName: \"kubernetes.io/projected/093c64a9-5cfc-4307-bc50-b8a50c263b48-kube-api-access-2m7fg\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:10 crc kubenswrapper[4733]: I1204 18:34:10.058749 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-utilities\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:10 crc kubenswrapper[4733]: I1204 18:34:10.058779 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-catalog-content\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:10 crc kubenswrapper[4733]: I1204 18:34:10.059249 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-utilities\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:10 crc kubenswrapper[4733]: I1204 18:34:10.059458 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-catalog-content\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:10 crc kubenswrapper[4733]: I1204 18:34:10.091728 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m7fg\" (UniqueName: \"kubernetes.io/projected/093c64a9-5cfc-4307-bc50-b8a50c263b48-kube-api-access-2m7fg\") pod \"community-operators-w978q\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:10 crc kubenswrapper[4733]: I1204 18:34:10.192450 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:10 crc kubenswrapper[4733]: I1204 18:34:10.463267 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w978q"] Dec 04 18:34:11 crc kubenswrapper[4733]: I1204 18:34:11.266954 4733 generic.go:334] "Generic (PLEG): container finished" podID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerID="7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44" exitCode=0 Dec 04 18:34:11 crc kubenswrapper[4733]: I1204 18:34:11.267025 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w978q" event={"ID":"093c64a9-5cfc-4307-bc50-b8a50c263b48","Type":"ContainerDied","Data":"7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44"} Dec 04 18:34:11 crc kubenswrapper[4733]: I1204 18:34:11.267066 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w978q" event={"ID":"093c64a9-5cfc-4307-bc50-b8a50c263b48","Type":"ContainerStarted","Data":"215a0594675aced07f09101fe1b58f28174a0caca34e74fed9310548c36fdb71"} Dec 04 18:34:12 crc kubenswrapper[4733]: I1204 18:34:12.290935 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w978q" event={"ID":"093c64a9-5cfc-4307-bc50-b8a50c263b48","Type":"ContainerStarted","Data":"9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed"} Dec 04 18:34:13 crc kubenswrapper[4733]: I1204 18:34:13.302267 4733 generic.go:334] "Generic (PLEG): container finished" podID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerID="9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed" exitCode=0 Dec 04 18:34:13 crc kubenswrapper[4733]: I1204 18:34:13.302334 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w978q" event={"ID":"093c64a9-5cfc-4307-bc50-b8a50c263b48","Type":"ContainerDied","Data":"9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed"} Dec 04 18:34:14 crc kubenswrapper[4733]: I1204 18:34:14.311231 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w978q" event={"ID":"093c64a9-5cfc-4307-bc50-b8a50c263b48","Type":"ContainerStarted","Data":"f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052"} Dec 04 18:34:14 crc kubenswrapper[4733]: I1204 18:34:14.334453 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w978q" podStartSLOduration=2.919861795 podStartE2EDuration="5.334430062s" podCreationTimestamp="2025-12-04 18:34:09 +0000 UTC" firstStartedPulling="2025-12-04 18:34:11.270491316 +0000 UTC m=+3313.225852392" lastFinishedPulling="2025-12-04 18:34:13.685059583 +0000 UTC m=+3315.640420659" observedRunningTime="2025-12-04 18:34:14.328621785 +0000 UTC m=+3316.283982861" watchObservedRunningTime="2025-12-04 18:34:14.334430062 +0000 UTC m=+3316.289791128" Dec 04 18:34:16 crc kubenswrapper[4733]: I1204 18:34:16.335491 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:34:17 crc kubenswrapper[4733]: I1204 18:34:17.339009 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"21674decaacdc3513aac16de9e2ba8f6177477028096e8539bef2260994d364a"} Dec 04 18:34:20 crc kubenswrapper[4733]: I1204 18:34:20.193315 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:20 crc kubenswrapper[4733]: I1204 18:34:20.194120 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:20 crc kubenswrapper[4733]: I1204 18:34:20.244769 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:20 crc kubenswrapper[4733]: I1204 18:34:20.433159 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:20 crc kubenswrapper[4733]: I1204 18:34:20.498341 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w978q"] Dec 04 18:34:22 crc kubenswrapper[4733]: I1204 18:34:22.379336 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w978q" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerName="registry-server" containerID="cri-o://f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052" gracePeriod=2 Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.362405 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.394374 4733 generic.go:334] "Generic (PLEG): container finished" podID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerID="f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052" exitCode=0 Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.394434 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w978q" event={"ID":"093c64a9-5cfc-4307-bc50-b8a50c263b48","Type":"ContainerDied","Data":"f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052"} Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.394504 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w978q" event={"ID":"093c64a9-5cfc-4307-bc50-b8a50c263b48","Type":"ContainerDied","Data":"215a0594675aced07f09101fe1b58f28174a0caca34e74fed9310548c36fdb71"} Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.394535 4733 scope.go:117] "RemoveContainer" containerID="f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.394755 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w978q" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.426734 4733 scope.go:117] "RemoveContainer" containerID="9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.452864 4733 scope.go:117] "RemoveContainer" containerID="7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.496736 4733 scope.go:117] "RemoveContainer" containerID="f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052" Dec 04 18:34:23 crc kubenswrapper[4733]: E1204 18:34:23.497658 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052\": container with ID starting with f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052 not found: ID does not exist" containerID="f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.497744 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052"} err="failed to get container status \"f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052\": rpc error: code = NotFound desc = could not find container \"f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052\": container with ID starting with f0297cce10627e9cdc402ffc95c69010eb333eb7f35a9f0761b15532d0d6f052 not found: ID does not exist" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.497853 4733 scope.go:117] "RemoveContainer" containerID="9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed" Dec 04 18:34:23 crc kubenswrapper[4733]: E1204 18:34:23.498749 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed\": container with ID starting with 9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed not found: ID does not exist" containerID="9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.498781 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed"} err="failed to get container status \"9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed\": rpc error: code = NotFound desc = could not find container \"9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed\": container with ID starting with 9fd4e469bafa54cf4d140fc90fcfbcca14ddd79da4b4cf61cc61a0ecf9590bed not found: ID does not exist" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.498815 4733 scope.go:117] "RemoveContainer" containerID="7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44" Dec 04 18:34:23 crc kubenswrapper[4733]: E1204 18:34:23.499268 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44\": container with ID starting with 7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44 not found: ID does not exist" containerID="7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.499301 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44"} err="failed to get container status \"7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44\": rpc error: code = NotFound desc = could not find container \"7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44\": container with ID starting with 7de6664b0240b2e6fe0297f8b5a164b232c2f8b65aadd8b153bdab37e92efc44 not found: ID does not exist" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.501853 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m7fg\" (UniqueName: \"kubernetes.io/projected/093c64a9-5cfc-4307-bc50-b8a50c263b48-kube-api-access-2m7fg\") pod \"093c64a9-5cfc-4307-bc50-b8a50c263b48\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.502004 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-utilities\") pod \"093c64a9-5cfc-4307-bc50-b8a50c263b48\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.503649 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-utilities" (OuterVolumeSpecName: "utilities") pod "093c64a9-5cfc-4307-bc50-b8a50c263b48" (UID: "093c64a9-5cfc-4307-bc50-b8a50c263b48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.503869 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-catalog-content\") pod \"093c64a9-5cfc-4307-bc50-b8a50c263b48\" (UID: \"093c64a9-5cfc-4307-bc50-b8a50c263b48\") " Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.507860 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093c64a9-5cfc-4307-bc50-b8a50c263b48-kube-api-access-2m7fg" (OuterVolumeSpecName: "kube-api-access-2m7fg") pod "093c64a9-5cfc-4307-bc50-b8a50c263b48" (UID: "093c64a9-5cfc-4307-bc50-b8a50c263b48"). InnerVolumeSpecName "kube-api-access-2m7fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.514517 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m7fg\" (UniqueName: \"kubernetes.io/projected/093c64a9-5cfc-4307-bc50-b8a50c263b48-kube-api-access-2m7fg\") on node \"crc\" DevicePath \"\"" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.514570 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.585732 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "093c64a9-5cfc-4307-bc50-b8a50c263b48" (UID: "093c64a9-5cfc-4307-bc50-b8a50c263b48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.615955 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093c64a9-5cfc-4307-bc50-b8a50c263b48-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.746279 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w978q"] Dec 04 18:34:23 crc kubenswrapper[4733]: I1204 18:34:23.754780 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w978q"] Dec 04 18:34:24 crc kubenswrapper[4733]: I1204 18:34:24.348748 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" path="/var/lib/kubelet/pods/093c64a9-5cfc-4307-bc50-b8a50c263b48/volumes" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.115033 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2mljr"] Dec 04 18:36:39 crc kubenswrapper[4733]: E1204 18:36:39.116149 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerName="extract-content" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.116180 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerName="extract-content" Dec 04 18:36:39 crc kubenswrapper[4733]: E1204 18:36:39.116232 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerName="registry-server" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.116245 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerName="registry-server" Dec 04 18:36:39 crc kubenswrapper[4733]: E1204 18:36:39.116290 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerName="extract-utilities" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.116304 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerName="extract-utilities" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.116624 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="093c64a9-5cfc-4307-bc50-b8a50c263b48" containerName="registry-server" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.118713 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.134525 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2mljr"] Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.316631 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6mxf\" (UniqueName: \"kubernetes.io/projected/65d0b9ad-6075-41a5-9880-8a34ac58bb88-kube-api-access-n6mxf\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.316734 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-utilities\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.316810 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-catalog-content\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.418341 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-utilities\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.419193 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-utilities\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.419407 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-catalog-content\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.419679 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6mxf\" (UniqueName: \"kubernetes.io/projected/65d0b9ad-6075-41a5-9880-8a34ac58bb88-kube-api-access-n6mxf\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.420205 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-catalog-content\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.447192 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6mxf\" (UniqueName: \"kubernetes.io/projected/65d0b9ad-6075-41a5-9880-8a34ac58bb88-kube-api-access-n6mxf\") pod \"redhat-operators-2mljr\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.481005 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:39 crc kubenswrapper[4733]: I1204 18:36:39.950859 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2mljr"] Dec 04 18:36:39 crc kubenswrapper[4733]: W1204 18:36:39.953729 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65d0b9ad_6075_41a5_9880_8a34ac58bb88.slice/crio-cddb0a93c785e16a793ae9d00ce17d8e3244c34796a5365cf74654a6a417a4c0 WatchSource:0}: Error finding container cddb0a93c785e16a793ae9d00ce17d8e3244c34796a5365cf74654a6a417a4c0: Status 404 returned error can't find the container with id cddb0a93c785e16a793ae9d00ce17d8e3244c34796a5365cf74654a6a417a4c0 Dec 04 18:36:40 crc kubenswrapper[4733]: I1204 18:36:40.539039 4733 generic.go:334] "Generic (PLEG): container finished" podID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerID="e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726" exitCode=0 Dec 04 18:36:40 crc kubenswrapper[4733]: I1204 18:36:40.539094 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mljr" event={"ID":"65d0b9ad-6075-41a5-9880-8a34ac58bb88","Type":"ContainerDied","Data":"e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726"} Dec 04 18:36:40 crc kubenswrapper[4733]: I1204 18:36:40.539125 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mljr" event={"ID":"65d0b9ad-6075-41a5-9880-8a34ac58bb88","Type":"ContainerStarted","Data":"cddb0a93c785e16a793ae9d00ce17d8e3244c34796a5365cf74654a6a417a4c0"} Dec 04 18:36:40 crc kubenswrapper[4733]: I1204 18:36:40.541742 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:36:41 crc kubenswrapper[4733]: I1204 18:36:41.552532 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mljr" event={"ID":"65d0b9ad-6075-41a5-9880-8a34ac58bb88","Type":"ContainerStarted","Data":"43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53"} Dec 04 18:36:42 crc kubenswrapper[4733]: I1204 18:36:42.564889 4733 generic.go:334] "Generic (PLEG): container finished" podID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerID="43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53" exitCode=0 Dec 04 18:36:42 crc kubenswrapper[4733]: I1204 18:36:42.564958 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mljr" event={"ID":"65d0b9ad-6075-41a5-9880-8a34ac58bb88","Type":"ContainerDied","Data":"43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53"} Dec 04 18:36:43 crc kubenswrapper[4733]: I1204 18:36:43.578610 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mljr" event={"ID":"65d0b9ad-6075-41a5-9880-8a34ac58bb88","Type":"ContainerStarted","Data":"e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b"} Dec 04 18:36:45 crc kubenswrapper[4733]: I1204 18:36:45.362023 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:36:45 crc kubenswrapper[4733]: I1204 18:36:45.363046 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:36:49 crc kubenswrapper[4733]: I1204 18:36:49.481207 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:49 crc kubenswrapper[4733]: I1204 18:36:49.482558 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:50 crc kubenswrapper[4733]: I1204 18:36:50.552977 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2mljr" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="registry-server" probeResult="failure" output=< Dec 04 18:36:50 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 18:36:50 crc kubenswrapper[4733]: > Dec 04 18:36:59 crc kubenswrapper[4733]: I1204 18:36:59.540551 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:59 crc kubenswrapper[4733]: I1204 18:36:59.579371 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2mljr" podStartSLOduration=18.138345601 podStartE2EDuration="20.579341574s" podCreationTimestamp="2025-12-04 18:36:39 +0000 UTC" firstStartedPulling="2025-12-04 18:36:40.541473232 +0000 UTC m=+3462.496834278" lastFinishedPulling="2025-12-04 18:36:42.982469175 +0000 UTC m=+3464.937830251" observedRunningTime="2025-12-04 18:36:43.614414961 +0000 UTC m=+3465.569776047" watchObservedRunningTime="2025-12-04 18:36:59.579341574 +0000 UTC m=+3481.534702660" Dec 04 18:36:59 crc kubenswrapper[4733]: I1204 18:36:59.624486 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:36:59 crc kubenswrapper[4733]: I1204 18:36:59.783334 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2mljr"] Dec 04 18:37:00 crc kubenswrapper[4733]: I1204 18:37:00.750080 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2mljr" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="registry-server" containerID="cri-o://e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b" gracePeriod=2 Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.241554 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.301707 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-utilities\") pod \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.301882 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6mxf\" (UniqueName: \"kubernetes.io/projected/65d0b9ad-6075-41a5-9880-8a34ac58bb88-kube-api-access-n6mxf\") pod \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.301904 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-catalog-content\") pod \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\" (UID: \"65d0b9ad-6075-41a5-9880-8a34ac58bb88\") " Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.302770 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-utilities" (OuterVolumeSpecName: "utilities") pod "65d0b9ad-6075-41a5-9880-8a34ac58bb88" (UID: "65d0b9ad-6075-41a5-9880-8a34ac58bb88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.307284 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d0b9ad-6075-41a5-9880-8a34ac58bb88-kube-api-access-n6mxf" (OuterVolumeSpecName: "kube-api-access-n6mxf") pod "65d0b9ad-6075-41a5-9880-8a34ac58bb88" (UID: "65d0b9ad-6075-41a5-9880-8a34ac58bb88"). InnerVolumeSpecName "kube-api-access-n6mxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.403516 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6mxf\" (UniqueName: \"kubernetes.io/projected/65d0b9ad-6075-41a5-9880-8a34ac58bb88-kube-api-access-n6mxf\") on node \"crc\" DevicePath \"\"" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.404122 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.408302 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65d0b9ad-6075-41a5-9880-8a34ac58bb88" (UID: "65d0b9ad-6075-41a5-9880-8a34ac58bb88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.506252 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d0b9ad-6075-41a5-9880-8a34ac58bb88-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.764726 4733 generic.go:334] "Generic (PLEG): container finished" podID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerID="e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b" exitCode=0 Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.764834 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mljr" event={"ID":"65d0b9ad-6075-41a5-9880-8a34ac58bb88","Type":"ContainerDied","Data":"e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b"} Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.764903 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mljr" event={"ID":"65d0b9ad-6075-41a5-9880-8a34ac58bb88","Type":"ContainerDied","Data":"cddb0a93c785e16a793ae9d00ce17d8e3244c34796a5365cf74654a6a417a4c0"} Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.764945 4733 scope.go:117] "RemoveContainer" containerID="e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.766686 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mljr" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.797642 4733 scope.go:117] "RemoveContainer" containerID="43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.819581 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2mljr"] Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.829208 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2mljr"] Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.835883 4733 scope.go:117] "RemoveContainer" containerID="e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.874358 4733 scope.go:117] "RemoveContainer" containerID="e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b" Dec 04 18:37:01 crc kubenswrapper[4733]: E1204 18:37:01.875407 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b\": container with ID starting with e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b not found: ID does not exist" containerID="e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.875476 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b"} err="failed to get container status \"e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b\": rpc error: code = NotFound desc = could not find container \"e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b\": container with ID starting with e46fc760c64da94acad58dc26f8e62d3e262e3c92f3606d5915aad8ebc411e3b not found: ID does not exist" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.875516 4733 scope.go:117] "RemoveContainer" containerID="43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53" Dec 04 18:37:01 crc kubenswrapper[4733]: E1204 18:37:01.876494 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53\": container with ID starting with 43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53 not found: ID does not exist" containerID="43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.876609 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53"} err="failed to get container status \"43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53\": rpc error: code = NotFound desc = could not find container \"43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53\": container with ID starting with 43e9e489d4343f3b0b8de36335382baeb908501b64698cfe6593363b334a1e53 not found: ID does not exist" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.876746 4733 scope.go:117] "RemoveContainer" containerID="e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726" Dec 04 18:37:01 crc kubenswrapper[4733]: E1204 18:37:01.877564 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726\": container with ID starting with e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726 not found: ID does not exist" containerID="e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726" Dec 04 18:37:01 crc kubenswrapper[4733]: I1204 18:37:01.877700 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726"} err="failed to get container status \"e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726\": rpc error: code = NotFound desc = could not find container \"e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726\": container with ID starting with e565d34a868acf7a456b3893ee96fa987047d12c8daa554d1363988060f0f726 not found: ID does not exist" Dec 04 18:37:02 crc kubenswrapper[4733]: I1204 18:37:02.352562 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" path="/var/lib/kubelet/pods/65d0b9ad-6075-41a5-9880-8a34ac58bb88/volumes" Dec 04 18:37:15 crc kubenswrapper[4733]: I1204 18:37:15.361938 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:37:15 crc kubenswrapper[4733]: I1204 18:37:15.362642 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.695601 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9xzjm"] Dec 04 18:37:26 crc kubenswrapper[4733]: E1204 18:37:26.698287 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="extract-utilities" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.698431 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="extract-utilities" Dec 04 18:37:26 crc kubenswrapper[4733]: E1204 18:37:26.698482 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="extract-content" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.698569 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="extract-content" Dec 04 18:37:26 crc kubenswrapper[4733]: E1204 18:37:26.698658 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="registry-server" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.698683 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="registry-server" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.699134 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d0b9ad-6075-41a5-9880-8a34ac58bb88" containerName="registry-server" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.701526 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.714124 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9xzjm"] Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.798625 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-utilities\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.799038 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-catalog-content\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.799281 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5fgc\" (UniqueName: \"kubernetes.io/projected/01a1f902-0369-410c-b464-9513631a2998-kube-api-access-d5fgc\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.900901 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-utilities\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.900983 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-catalog-content\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.901225 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5fgc\" (UniqueName: \"kubernetes.io/projected/01a1f902-0369-410c-b464-9513631a2998-kube-api-access-d5fgc\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.901652 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-utilities\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.901669 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-catalog-content\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:26 crc kubenswrapper[4733]: I1204 18:37:26.931195 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5fgc\" (UniqueName: \"kubernetes.io/projected/01a1f902-0369-410c-b464-9513631a2998-kube-api-access-d5fgc\") pod \"certified-operators-9xzjm\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:27 crc kubenswrapper[4733]: I1204 18:37:27.071380 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:27 crc kubenswrapper[4733]: I1204 18:37:27.521850 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9xzjm"] Dec 04 18:37:28 crc kubenswrapper[4733]: I1204 18:37:28.020398 4733 generic.go:334] "Generic (PLEG): container finished" podID="01a1f902-0369-410c-b464-9513631a2998" containerID="8e16ec02607a3142aa831957bcba73fc41fbdcce7f0f2f76121d7e2d14759a9c" exitCode=0 Dec 04 18:37:28 crc kubenswrapper[4733]: I1204 18:37:28.020573 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xzjm" event={"ID":"01a1f902-0369-410c-b464-9513631a2998","Type":"ContainerDied","Data":"8e16ec02607a3142aa831957bcba73fc41fbdcce7f0f2f76121d7e2d14759a9c"} Dec 04 18:37:28 crc kubenswrapper[4733]: I1204 18:37:28.020617 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xzjm" event={"ID":"01a1f902-0369-410c-b464-9513631a2998","Type":"ContainerStarted","Data":"416ad4f1e9fbce0292be19529f496775b9a2a7ac47c2e790dde680818a15ef5d"} Dec 04 18:37:30 crc kubenswrapper[4733]: I1204 18:37:30.040544 4733 generic.go:334] "Generic (PLEG): container finished" podID="01a1f902-0369-410c-b464-9513631a2998" containerID="9a0e61790e5634573364cd40a358813d438041dfd355289bd76f9175eac9ad76" exitCode=0 Dec 04 18:37:30 crc kubenswrapper[4733]: I1204 18:37:30.040615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xzjm" event={"ID":"01a1f902-0369-410c-b464-9513631a2998","Type":"ContainerDied","Data":"9a0e61790e5634573364cd40a358813d438041dfd355289bd76f9175eac9ad76"} Dec 04 18:37:31 crc kubenswrapper[4733]: I1204 18:37:31.051741 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xzjm" event={"ID":"01a1f902-0369-410c-b464-9513631a2998","Type":"ContainerStarted","Data":"aea816fa5ec0505513f3ad27747e00a6bb2262485ab0d156a33d0bb680960e23"} Dec 04 18:37:31 crc kubenswrapper[4733]: I1204 18:37:31.072354 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9xzjm" podStartSLOduration=2.6434467550000003 podStartE2EDuration="5.07233461s" podCreationTimestamp="2025-12-04 18:37:26 +0000 UTC" firstStartedPulling="2025-12-04 18:37:28.024389858 +0000 UTC m=+3509.979750944" lastFinishedPulling="2025-12-04 18:37:30.453277713 +0000 UTC m=+3512.408638799" observedRunningTime="2025-12-04 18:37:31.068561408 +0000 UTC m=+3513.023922454" watchObservedRunningTime="2025-12-04 18:37:31.07233461 +0000 UTC m=+3513.027695656" Dec 04 18:37:37 crc kubenswrapper[4733]: I1204 18:37:37.072167 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:37 crc kubenswrapper[4733]: I1204 18:37:37.072850 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:37 crc kubenswrapper[4733]: I1204 18:37:37.146093 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:37 crc kubenswrapper[4733]: I1204 18:37:37.228246 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:37 crc kubenswrapper[4733]: I1204 18:37:37.402419 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9xzjm"] Dec 04 18:37:39 crc kubenswrapper[4733]: I1204 18:37:39.128189 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9xzjm" podUID="01a1f902-0369-410c-b464-9513631a2998" containerName="registry-server" containerID="cri-o://aea816fa5ec0505513f3ad27747e00a6bb2262485ab0d156a33d0bb680960e23" gracePeriod=2 Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.139141 4733 generic.go:334] "Generic (PLEG): container finished" podID="01a1f902-0369-410c-b464-9513631a2998" containerID="aea816fa5ec0505513f3ad27747e00a6bb2262485ab0d156a33d0bb680960e23" exitCode=0 Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.139226 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xzjm" event={"ID":"01a1f902-0369-410c-b464-9513631a2998","Type":"ContainerDied","Data":"aea816fa5ec0505513f3ad27747e00a6bb2262485ab0d156a33d0bb680960e23"} Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.760901 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.919005 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5fgc\" (UniqueName: \"kubernetes.io/projected/01a1f902-0369-410c-b464-9513631a2998-kube-api-access-d5fgc\") pod \"01a1f902-0369-410c-b464-9513631a2998\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.919575 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-catalog-content\") pod \"01a1f902-0369-410c-b464-9513631a2998\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.919683 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-utilities\") pod \"01a1f902-0369-410c-b464-9513631a2998\" (UID: \"01a1f902-0369-410c-b464-9513631a2998\") " Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.920652 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-utilities" (OuterVolumeSpecName: "utilities") pod "01a1f902-0369-410c-b464-9513631a2998" (UID: "01a1f902-0369-410c-b464-9513631a2998"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.926319 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a1f902-0369-410c-b464-9513631a2998-kube-api-access-d5fgc" (OuterVolumeSpecName: "kube-api-access-d5fgc") pod "01a1f902-0369-410c-b464-9513631a2998" (UID: "01a1f902-0369-410c-b464-9513631a2998"). InnerVolumeSpecName "kube-api-access-d5fgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:37:40 crc kubenswrapper[4733]: I1204 18:37:40.986988 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01a1f902-0369-410c-b464-9513631a2998" (UID: "01a1f902-0369-410c-b464-9513631a2998"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.021429 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5fgc\" (UniqueName: \"kubernetes.io/projected/01a1f902-0369-410c-b464-9513631a2998-kube-api-access-d5fgc\") on node \"crc\" DevicePath \"\"" Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.021467 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.021482 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a1f902-0369-410c-b464-9513631a2998-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.153265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xzjm" event={"ID":"01a1f902-0369-410c-b464-9513631a2998","Type":"ContainerDied","Data":"416ad4f1e9fbce0292be19529f496775b9a2a7ac47c2e790dde680818a15ef5d"} Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.153330 4733 scope.go:117] "RemoveContainer" containerID="aea816fa5ec0505513f3ad27747e00a6bb2262485ab0d156a33d0bb680960e23" Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.153334 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xzjm" Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.184445 4733 scope.go:117] "RemoveContainer" containerID="9a0e61790e5634573364cd40a358813d438041dfd355289bd76f9175eac9ad76" Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.201285 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9xzjm"] Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.208227 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9xzjm"] Dec 04 18:37:41 crc kubenswrapper[4733]: I1204 18:37:41.222490 4733 scope.go:117] "RemoveContainer" containerID="8e16ec02607a3142aa831957bcba73fc41fbdcce7f0f2f76121d7e2d14759a9c" Dec 04 18:37:42 crc kubenswrapper[4733]: I1204 18:37:42.352754 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a1f902-0369-410c-b464-9513631a2998" path="/var/lib/kubelet/pods/01a1f902-0369-410c-b464-9513631a2998/volumes" Dec 04 18:37:45 crc kubenswrapper[4733]: I1204 18:37:45.362465 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:37:45 crc kubenswrapper[4733]: I1204 18:37:45.363255 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:37:45 crc kubenswrapper[4733]: I1204 18:37:45.363354 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:37:45 crc kubenswrapper[4733]: I1204 18:37:45.364598 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21674decaacdc3513aac16de9e2ba8f6177477028096e8539bef2260994d364a"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:37:45 crc kubenswrapper[4733]: I1204 18:37:45.364745 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://21674decaacdc3513aac16de9e2ba8f6177477028096e8539bef2260994d364a" gracePeriod=600 Dec 04 18:37:46 crc kubenswrapper[4733]: I1204 18:37:46.204726 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="21674decaacdc3513aac16de9e2ba8f6177477028096e8539bef2260994d364a" exitCode=0 Dec 04 18:37:46 crc kubenswrapper[4733]: I1204 18:37:46.204790 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"21674decaacdc3513aac16de9e2ba8f6177477028096e8539bef2260994d364a"} Dec 04 18:37:46 crc kubenswrapper[4733]: I1204 18:37:46.205284 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71"} Dec 04 18:37:46 crc kubenswrapper[4733]: I1204 18:37:46.205334 4733 scope.go:117] "RemoveContainer" containerID="9ce796731942f4c067b039dad573a00ab055ea68c67f005f803fbe16545ba759" Dec 04 18:39:45 crc kubenswrapper[4733]: I1204 18:39:45.362487 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:39:45 crc kubenswrapper[4733]: I1204 18:39:45.363116 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:40:15 crc kubenswrapper[4733]: I1204 18:40:15.362325 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:40:15 crc kubenswrapper[4733]: I1204 18:40:15.363217 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.362080 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.362790 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.362870 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.363472 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.363564 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" gracePeriod=600 Dec 04 18:40:45 crc kubenswrapper[4733]: E1204 18:40:45.522662 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.920849 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" exitCode=0 Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.920920 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71"} Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.920976 4733 scope.go:117] "RemoveContainer" containerID="21674decaacdc3513aac16de9e2ba8f6177477028096e8539bef2260994d364a" Dec 04 18:40:45 crc kubenswrapper[4733]: I1204 18:40:45.921585 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:40:45 crc kubenswrapper[4733]: E1204 18:40:45.922163 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:41:00 crc kubenswrapper[4733]: I1204 18:41:00.337114 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:41:00 crc kubenswrapper[4733]: E1204 18:41:00.338184 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:41:11 crc kubenswrapper[4733]: I1204 18:41:11.336677 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:41:11 crc kubenswrapper[4733]: E1204 18:41:11.340546 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.037678 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-597m5"] Dec 04 18:41:15 crc kubenswrapper[4733]: E1204 18:41:15.038595 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a1f902-0369-410c-b464-9513631a2998" containerName="extract-utilities" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.038628 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a1f902-0369-410c-b464-9513631a2998" containerName="extract-utilities" Dec 04 18:41:15 crc kubenswrapper[4733]: E1204 18:41:15.038659 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a1f902-0369-410c-b464-9513631a2998" containerName="registry-server" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.038679 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a1f902-0369-410c-b464-9513631a2998" containerName="registry-server" Dec 04 18:41:15 crc kubenswrapper[4733]: E1204 18:41:15.038732 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a1f902-0369-410c-b464-9513631a2998" containerName="extract-content" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.038752 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a1f902-0369-410c-b464-9513631a2998" containerName="extract-content" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.039343 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a1f902-0369-410c-b464-9513631a2998" containerName="registry-server" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.041647 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.070939 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-597m5"] Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.208658 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksrhm\" (UniqueName: \"kubernetes.io/projected/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-kube-api-access-ksrhm\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.208760 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-catalog-content\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.208823 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-utilities\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.310293 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-utilities\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.310413 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksrhm\" (UniqueName: \"kubernetes.io/projected/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-kube-api-access-ksrhm\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.310514 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-catalog-content\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.311170 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-catalog-content\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.311497 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-utilities\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.345035 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksrhm\" (UniqueName: \"kubernetes.io/projected/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-kube-api-access-ksrhm\") pod \"redhat-marketplace-597m5\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.370857 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:15 crc kubenswrapper[4733]: I1204 18:41:15.646313 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-597m5"] Dec 04 18:41:16 crc kubenswrapper[4733]: I1204 18:41:16.238769 4733 generic.go:334] "Generic (PLEG): container finished" podID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerID="e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2" exitCode=0 Dec 04 18:41:16 crc kubenswrapper[4733]: I1204 18:41:16.238832 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-597m5" event={"ID":"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a","Type":"ContainerDied","Data":"e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2"} Dec 04 18:41:16 crc kubenswrapper[4733]: I1204 18:41:16.239212 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-597m5" event={"ID":"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a","Type":"ContainerStarted","Data":"34bd3fdccb25b5795a59338f7b24cf547061880dce03d4c5f07c4778092ecf01"} Dec 04 18:41:18 crc kubenswrapper[4733]: I1204 18:41:18.264174 4733 generic.go:334] "Generic (PLEG): container finished" podID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerID="50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c" exitCode=0 Dec 04 18:41:18 crc kubenswrapper[4733]: I1204 18:41:18.264248 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-597m5" event={"ID":"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a","Type":"ContainerDied","Data":"50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c"} Dec 04 18:41:19 crc kubenswrapper[4733]: I1204 18:41:19.276612 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-597m5" event={"ID":"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a","Type":"ContainerStarted","Data":"ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552"} Dec 04 18:41:24 crc kubenswrapper[4733]: I1204 18:41:24.344263 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:41:24 crc kubenswrapper[4733]: E1204 18:41:24.345378 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:41:25 crc kubenswrapper[4733]: I1204 18:41:25.371839 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:25 crc kubenswrapper[4733]: I1204 18:41:25.371923 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:25 crc kubenswrapper[4733]: I1204 18:41:25.456812 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:25 crc kubenswrapper[4733]: I1204 18:41:25.478742 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-597m5" podStartSLOduration=7.818691555 podStartE2EDuration="10.47872718s" podCreationTimestamp="2025-12-04 18:41:15 +0000 UTC" firstStartedPulling="2025-12-04 18:41:16.240521107 +0000 UTC m=+3738.195882153" lastFinishedPulling="2025-12-04 18:41:18.900556722 +0000 UTC m=+3740.855917778" observedRunningTime="2025-12-04 18:41:19.306257425 +0000 UTC m=+3741.261618541" watchObservedRunningTime="2025-12-04 18:41:25.47872718 +0000 UTC m=+3747.434088226" Dec 04 18:41:26 crc kubenswrapper[4733]: I1204 18:41:26.404345 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:26 crc kubenswrapper[4733]: I1204 18:41:26.452450 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-597m5"] Dec 04 18:41:28 crc kubenswrapper[4733]: I1204 18:41:28.364421 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-597m5" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerName="registry-server" containerID="cri-o://ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552" gracePeriod=2 Dec 04 18:41:28 crc kubenswrapper[4733]: I1204 18:41:28.845292 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:28 crc kubenswrapper[4733]: I1204 18:41:28.931404 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-catalog-content\") pod \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " Dec 04 18:41:28 crc kubenswrapper[4733]: I1204 18:41:28.931480 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-utilities\") pod \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " Dec 04 18:41:28 crc kubenswrapper[4733]: I1204 18:41:28.931581 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksrhm\" (UniqueName: \"kubernetes.io/projected/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-kube-api-access-ksrhm\") pod \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\" (UID: \"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a\") " Dec 04 18:41:28 crc kubenswrapper[4733]: I1204 18:41:28.932443 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-utilities" (OuterVolumeSpecName: "utilities") pod "70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" (UID: "70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:41:28 crc kubenswrapper[4733]: I1204 18:41:28.938160 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-kube-api-access-ksrhm" (OuterVolumeSpecName: "kube-api-access-ksrhm") pod "70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" (UID: "70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a"). InnerVolumeSpecName "kube-api-access-ksrhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:41:28 crc kubenswrapper[4733]: I1204 18:41:28.955727 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" (UID: "70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.033611 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksrhm\" (UniqueName: \"kubernetes.io/projected/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-kube-api-access-ksrhm\") on node \"crc\" DevicePath \"\"" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.033670 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.033699 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.378332 4733 generic.go:334] "Generic (PLEG): container finished" podID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerID="ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552" exitCode=0 Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.378404 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-597m5" event={"ID":"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a","Type":"ContainerDied","Data":"ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552"} Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.378446 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-597m5" event={"ID":"70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a","Type":"ContainerDied","Data":"34bd3fdccb25b5795a59338f7b24cf547061880dce03d4c5f07c4778092ecf01"} Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.378445 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-597m5" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.378474 4733 scope.go:117] "RemoveContainer" containerID="ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.413160 4733 scope.go:117] "RemoveContainer" containerID="50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.428927 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-597m5"] Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.435781 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-597m5"] Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.452770 4733 scope.go:117] "RemoveContainer" containerID="e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.497894 4733 scope.go:117] "RemoveContainer" containerID="ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552" Dec 04 18:41:29 crc kubenswrapper[4733]: E1204 18:41:29.498363 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552\": container with ID starting with ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552 not found: ID does not exist" containerID="ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.498438 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552"} err="failed to get container status \"ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552\": rpc error: code = NotFound desc = could not find container \"ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552\": container with ID starting with ee0721e36f56781271cd9ca3472aec2460f7c52dd09e02879d156ff8d6e12552 not found: ID does not exist" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.498488 4733 scope.go:117] "RemoveContainer" containerID="50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c" Dec 04 18:41:29 crc kubenswrapper[4733]: E1204 18:41:29.499204 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c\": container with ID starting with 50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c not found: ID does not exist" containerID="50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.499258 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c"} err="failed to get container status \"50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c\": rpc error: code = NotFound desc = could not find container \"50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c\": container with ID starting with 50ab64b64ef0d0b9c5b3c0880590a84abca259d327dfd451337db23bcab1857c not found: ID does not exist" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.499291 4733 scope.go:117] "RemoveContainer" containerID="e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2" Dec 04 18:41:29 crc kubenswrapper[4733]: E1204 18:41:29.499646 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2\": container with ID starting with e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2 not found: ID does not exist" containerID="e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2" Dec 04 18:41:29 crc kubenswrapper[4733]: I1204 18:41:29.499696 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2"} err="failed to get container status \"e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2\": rpc error: code = NotFound desc = could not find container \"e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2\": container with ID starting with e72059fefbe0070a0c95061a668f5622de62953f085fea90fa4e2b0a98844dc2 not found: ID does not exist" Dec 04 18:41:30 crc kubenswrapper[4733]: I1204 18:41:30.353285 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" path="/var/lib/kubelet/pods/70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a/volumes" Dec 04 18:41:39 crc kubenswrapper[4733]: I1204 18:41:39.335507 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:41:39 crc kubenswrapper[4733]: E1204 18:41:39.336648 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:41:51 crc kubenswrapper[4733]: I1204 18:41:51.335571 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:41:51 crc kubenswrapper[4733]: E1204 18:41:51.336732 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:42:05 crc kubenswrapper[4733]: I1204 18:42:05.335321 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:42:05 crc kubenswrapper[4733]: E1204 18:42:05.335874 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:42:20 crc kubenswrapper[4733]: I1204 18:42:20.335016 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:42:20 crc kubenswrapper[4733]: E1204 18:42:20.335975 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:42:32 crc kubenswrapper[4733]: I1204 18:42:32.335881 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:42:32 crc kubenswrapper[4733]: E1204 18:42:32.336905 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:42:45 crc kubenswrapper[4733]: I1204 18:42:45.336197 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:42:45 crc kubenswrapper[4733]: E1204 18:42:45.337718 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:42:57 crc kubenswrapper[4733]: I1204 18:42:57.336075 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:42:57 crc kubenswrapper[4733]: E1204 18:42:57.337434 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:43:11 crc kubenswrapper[4733]: I1204 18:43:11.335969 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:43:11 crc kubenswrapper[4733]: E1204 18:43:11.337117 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:43:22 crc kubenswrapper[4733]: I1204 18:43:22.336453 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:43:22 crc kubenswrapper[4733]: E1204 18:43:22.337359 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:43:33 crc kubenswrapper[4733]: I1204 18:43:33.335430 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:43:33 crc kubenswrapper[4733]: E1204 18:43:33.336311 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:43:45 crc kubenswrapper[4733]: I1204 18:43:45.336397 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:43:45 crc kubenswrapper[4733]: E1204 18:43:45.337160 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:43:56 crc kubenswrapper[4733]: I1204 18:43:56.337502 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:43:56 crc kubenswrapper[4733]: E1204 18:43:56.338283 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:44:09 crc kubenswrapper[4733]: I1204 18:44:09.336470 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:44:09 crc kubenswrapper[4733]: E1204 18:44:09.337546 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:44:23 crc kubenswrapper[4733]: I1204 18:44:23.335860 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:44:23 crc kubenswrapper[4733]: E1204 18:44:23.336647 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:44:34 crc kubenswrapper[4733]: I1204 18:44:34.335426 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:44:34 crc kubenswrapper[4733]: E1204 18:44:34.337361 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.455270 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-26vs2"] Dec 04 18:44:47 crc kubenswrapper[4733]: E1204 18:44:47.457253 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerName="registry-server" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.457363 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerName="registry-server" Dec 04 18:44:47 crc kubenswrapper[4733]: E1204 18:44:47.457460 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerName="extract-content" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.457545 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerName="extract-content" Dec 04 18:44:47 crc kubenswrapper[4733]: E1204 18:44:47.457650 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerName="extract-utilities" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.457733 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerName="extract-utilities" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.458020 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="70b56bc4-a176-4bdc-9f07-0d5ddf16ab5a" containerName="registry-server" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.459427 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.462471 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-26vs2"] Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.524385 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-catalog-content\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.524427 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkpd8\" (UniqueName: \"kubernetes.io/projected/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-kube-api-access-tkpd8\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.524712 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-utilities\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.626101 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-utilities\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.626239 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-catalog-content\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.626287 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkpd8\" (UniqueName: \"kubernetes.io/projected/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-kube-api-access-tkpd8\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.626754 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-utilities\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.626871 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-catalog-content\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.663017 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkpd8\" (UniqueName: \"kubernetes.io/projected/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-kube-api-access-tkpd8\") pod \"community-operators-26vs2\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:47 crc kubenswrapper[4733]: I1204 18:44:47.784964 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:48 crc kubenswrapper[4733]: I1204 18:44:48.333637 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-26vs2"] Dec 04 18:44:48 crc kubenswrapper[4733]: I1204 18:44:48.342642 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:44:48 crc kubenswrapper[4733]: E1204 18:44:48.342843 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:44:48 crc kubenswrapper[4733]: W1204 18:44:48.350986 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7746e49f_8f15_4abd_938c_2aa5e2bc64fa.slice/crio-c2a903bc7de70f7300bf263e977e743e5fa9f6ab88b4d424696d320c3380ee75 WatchSource:0}: Error finding container c2a903bc7de70f7300bf263e977e743e5fa9f6ab88b4d424696d320c3380ee75: Status 404 returned error can't find the container with id c2a903bc7de70f7300bf263e977e743e5fa9f6ab88b4d424696d320c3380ee75 Dec 04 18:44:49 crc kubenswrapper[4733]: I1204 18:44:49.153835 4733 generic.go:334] "Generic (PLEG): container finished" podID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerID="d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9" exitCode=0 Dec 04 18:44:49 crc kubenswrapper[4733]: I1204 18:44:49.153894 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26vs2" event={"ID":"7746e49f-8f15-4abd-938c-2aa5e2bc64fa","Type":"ContainerDied","Data":"d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9"} Dec 04 18:44:49 crc kubenswrapper[4733]: I1204 18:44:49.154160 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26vs2" event={"ID":"7746e49f-8f15-4abd-938c-2aa5e2bc64fa","Type":"ContainerStarted","Data":"c2a903bc7de70f7300bf263e977e743e5fa9f6ab88b4d424696d320c3380ee75"} Dec 04 18:44:49 crc kubenswrapper[4733]: I1204 18:44:49.156133 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:44:51 crc kubenswrapper[4733]: I1204 18:44:51.169847 4733 generic.go:334] "Generic (PLEG): container finished" podID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerID="95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25" exitCode=0 Dec 04 18:44:51 crc kubenswrapper[4733]: I1204 18:44:51.169974 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26vs2" event={"ID":"7746e49f-8f15-4abd-938c-2aa5e2bc64fa","Type":"ContainerDied","Data":"95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25"} Dec 04 18:44:52 crc kubenswrapper[4733]: I1204 18:44:52.179950 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26vs2" event={"ID":"7746e49f-8f15-4abd-938c-2aa5e2bc64fa","Type":"ContainerStarted","Data":"9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086"} Dec 04 18:44:52 crc kubenswrapper[4733]: I1204 18:44:52.204041 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-26vs2" podStartSLOduration=2.789216332 podStartE2EDuration="5.204024958s" podCreationTimestamp="2025-12-04 18:44:47 +0000 UTC" firstStartedPulling="2025-12-04 18:44:49.155878068 +0000 UTC m=+3951.111239124" lastFinishedPulling="2025-12-04 18:44:51.570686694 +0000 UTC m=+3953.526047750" observedRunningTime="2025-12-04 18:44:52.20373593 +0000 UTC m=+3954.159096986" watchObservedRunningTime="2025-12-04 18:44:52.204024958 +0000 UTC m=+3954.159386004" Dec 04 18:44:57 crc kubenswrapper[4733]: I1204 18:44:57.786180 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:57 crc kubenswrapper[4733]: I1204 18:44:57.786740 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:57 crc kubenswrapper[4733]: I1204 18:44:57.828131 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:58 crc kubenswrapper[4733]: I1204 18:44:58.301877 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:44:58 crc kubenswrapper[4733]: I1204 18:44:58.364633 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-26vs2"] Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.185339 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv"] Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.186927 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.190512 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.193784 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.196094 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv"] Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.210456 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7bfcbd43-0013-4715-81ad-dd378598b7e1-config-volume\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.210512 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7bfcbd43-0013-4715-81ad-dd378598b7e1-secret-volume\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.210561 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnvl8\" (UniqueName: \"kubernetes.io/projected/7bfcbd43-0013-4715-81ad-dd378598b7e1-kube-api-access-lnvl8\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.239744 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-26vs2" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerName="registry-server" containerID="cri-o://9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086" gracePeriod=2 Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.312276 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7bfcbd43-0013-4715-81ad-dd378598b7e1-config-volume\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.312356 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7bfcbd43-0013-4715-81ad-dd378598b7e1-secret-volume\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.312408 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnvl8\" (UniqueName: \"kubernetes.io/projected/7bfcbd43-0013-4715-81ad-dd378598b7e1-kube-api-access-lnvl8\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.313778 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7bfcbd43-0013-4715-81ad-dd378598b7e1-config-volume\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.318653 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7bfcbd43-0013-4715-81ad-dd378598b7e1-secret-volume\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.330530 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnvl8\" (UniqueName: \"kubernetes.io/projected/7bfcbd43-0013-4715-81ad-dd378598b7e1-kube-api-access-lnvl8\") pod \"collect-profiles-29414565-kd7vv\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.335891 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:45:00 crc kubenswrapper[4733]: E1204 18:45:00.336120 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.503556 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.765354 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.822093 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-utilities\") pod \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.822158 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-catalog-content\") pod \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.822244 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkpd8\" (UniqueName: \"kubernetes.io/projected/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-kube-api-access-tkpd8\") pod \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\" (UID: \"7746e49f-8f15-4abd-938c-2aa5e2bc64fa\") " Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.823146 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-utilities" (OuterVolumeSpecName: "utilities") pod "7746e49f-8f15-4abd-938c-2aa5e2bc64fa" (UID: "7746e49f-8f15-4abd-938c-2aa5e2bc64fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.826668 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-kube-api-access-tkpd8" (OuterVolumeSpecName: "kube-api-access-tkpd8") pod "7746e49f-8f15-4abd-938c-2aa5e2bc64fa" (UID: "7746e49f-8f15-4abd-938c-2aa5e2bc64fa"). InnerVolumeSpecName "kube-api-access-tkpd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.924072 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkpd8\" (UniqueName: \"kubernetes.io/projected/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-kube-api-access-tkpd8\") on node \"crc\" DevicePath \"\"" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.924110 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:45:00 crc kubenswrapper[4733]: I1204 18:45:00.964702 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7746e49f-8f15-4abd-938c-2aa5e2bc64fa" (UID: "7746e49f-8f15-4abd-938c-2aa5e2bc64fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.022491 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv"] Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.025916 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7746e49f-8f15-4abd-938c-2aa5e2bc64fa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:45:01 crc kubenswrapper[4733]: W1204 18:45:01.149157 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bfcbd43_0013_4715_81ad_dd378598b7e1.slice/crio-83021cd6dd084ffa1f27202d0d975319a87e0f543a7a07e111f737234fad891c WatchSource:0}: Error finding container 83021cd6dd084ffa1f27202d0d975319a87e0f543a7a07e111f737234fad891c: Status 404 returned error can't find the container with id 83021cd6dd084ffa1f27202d0d975319a87e0f543a7a07e111f737234fad891c Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.251181 4733 generic.go:334] "Generic (PLEG): container finished" podID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerID="9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086" exitCode=0 Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.251230 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26vs2" event={"ID":"7746e49f-8f15-4abd-938c-2aa5e2bc64fa","Type":"ContainerDied","Data":"9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086"} Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.251619 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26vs2" event={"ID":"7746e49f-8f15-4abd-938c-2aa5e2bc64fa","Type":"ContainerDied","Data":"c2a903bc7de70f7300bf263e977e743e5fa9f6ab88b4d424696d320c3380ee75"} Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.251647 4733 scope.go:117] "RemoveContainer" containerID="9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.251277 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-26vs2" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.255000 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" event={"ID":"7bfcbd43-0013-4715-81ad-dd378598b7e1","Type":"ContainerStarted","Data":"83021cd6dd084ffa1f27202d0d975319a87e0f543a7a07e111f737234fad891c"} Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.270288 4733 scope.go:117] "RemoveContainer" containerID="95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.288826 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-26vs2"] Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.296503 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-26vs2"] Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.310150 4733 scope.go:117] "RemoveContainer" containerID="d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.329221 4733 scope.go:117] "RemoveContainer" containerID="9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086" Dec 04 18:45:01 crc kubenswrapper[4733]: E1204 18:45:01.329673 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086\": container with ID starting with 9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086 not found: ID does not exist" containerID="9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.329711 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086"} err="failed to get container status \"9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086\": rpc error: code = NotFound desc = could not find container \"9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086\": container with ID starting with 9434e4181c7cdffb8f32edec982f4e21144cf56c98eb3a87a74f43ce065e1086 not found: ID does not exist" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.329739 4733 scope.go:117] "RemoveContainer" containerID="95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25" Dec 04 18:45:01 crc kubenswrapper[4733]: E1204 18:45:01.330224 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25\": container with ID starting with 95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25 not found: ID does not exist" containerID="95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.330251 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25"} err="failed to get container status \"95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25\": rpc error: code = NotFound desc = could not find container \"95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25\": container with ID starting with 95adaa91576ccee3b864e759a3ae93d775b0c04adbc95ad52d6011d1982dad25 not found: ID does not exist" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.330275 4733 scope.go:117] "RemoveContainer" containerID="d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9" Dec 04 18:45:01 crc kubenswrapper[4733]: E1204 18:45:01.330679 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9\": container with ID starting with d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9 not found: ID does not exist" containerID="d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9" Dec 04 18:45:01 crc kubenswrapper[4733]: I1204 18:45:01.330700 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9"} err="failed to get container status \"d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9\": rpc error: code = NotFound desc = could not find container \"d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9\": container with ID starting with d15b0f384ce07e0ec31106ad89bc6f2258e666eb549002d6cae3964bff8e14b9 not found: ID does not exist" Dec 04 18:45:02 crc kubenswrapper[4733]: I1204 18:45:02.263926 4733 generic.go:334] "Generic (PLEG): container finished" podID="7bfcbd43-0013-4715-81ad-dd378598b7e1" containerID="af7127c73e9826cf31c7d42cd1e66a5bdbccaf09d08eba57f9d1df90daee5456" exitCode=0 Dec 04 18:45:02 crc kubenswrapper[4733]: I1204 18:45:02.263982 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" event={"ID":"7bfcbd43-0013-4715-81ad-dd378598b7e1","Type":"ContainerDied","Data":"af7127c73e9826cf31c7d42cd1e66a5bdbccaf09d08eba57f9d1df90daee5456"} Dec 04 18:45:02 crc kubenswrapper[4733]: I1204 18:45:02.347340 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" path="/var/lib/kubelet/pods/7746e49f-8f15-4abd-938c-2aa5e2bc64fa/volumes" Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.718656 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.766653 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7bfcbd43-0013-4715-81ad-dd378598b7e1-config-volume\") pod \"7bfcbd43-0013-4715-81ad-dd378598b7e1\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.766723 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnvl8\" (UniqueName: \"kubernetes.io/projected/7bfcbd43-0013-4715-81ad-dd378598b7e1-kube-api-access-lnvl8\") pod \"7bfcbd43-0013-4715-81ad-dd378598b7e1\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.766747 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7bfcbd43-0013-4715-81ad-dd378598b7e1-secret-volume\") pod \"7bfcbd43-0013-4715-81ad-dd378598b7e1\" (UID: \"7bfcbd43-0013-4715-81ad-dd378598b7e1\") " Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.767287 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bfcbd43-0013-4715-81ad-dd378598b7e1-config-volume" (OuterVolumeSpecName: "config-volume") pod "7bfcbd43-0013-4715-81ad-dd378598b7e1" (UID: "7bfcbd43-0013-4715-81ad-dd378598b7e1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.777017 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bfcbd43-0013-4715-81ad-dd378598b7e1-kube-api-access-lnvl8" (OuterVolumeSpecName: "kube-api-access-lnvl8") pod "7bfcbd43-0013-4715-81ad-dd378598b7e1" (UID: "7bfcbd43-0013-4715-81ad-dd378598b7e1"). InnerVolumeSpecName "kube-api-access-lnvl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.791580 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfcbd43-0013-4715-81ad-dd378598b7e1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7bfcbd43-0013-4715-81ad-dd378598b7e1" (UID: "7bfcbd43-0013-4715-81ad-dd378598b7e1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.867873 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7bfcbd43-0013-4715-81ad-dd378598b7e1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.867909 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnvl8\" (UniqueName: \"kubernetes.io/projected/7bfcbd43-0013-4715-81ad-dd378598b7e1-kube-api-access-lnvl8\") on node \"crc\" DevicePath \"\"" Dec 04 18:45:03 crc kubenswrapper[4733]: I1204 18:45:03.867923 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7bfcbd43-0013-4715-81ad-dd378598b7e1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 18:45:04 crc kubenswrapper[4733]: I1204 18:45:04.283709 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" event={"ID":"7bfcbd43-0013-4715-81ad-dd378598b7e1","Type":"ContainerDied","Data":"83021cd6dd084ffa1f27202d0d975319a87e0f543a7a07e111f737234fad891c"} Dec 04 18:45:04 crc kubenswrapper[4733]: I1204 18:45:04.283784 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83021cd6dd084ffa1f27202d0d975319a87e0f543a7a07e111f737234fad891c" Dec 04 18:45:04 crc kubenswrapper[4733]: I1204 18:45:04.283809 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv" Dec 04 18:45:04 crc kubenswrapper[4733]: I1204 18:45:04.790134 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml"] Dec 04 18:45:04 crc kubenswrapper[4733]: I1204 18:45:04.796140 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414520-nzxml"] Dec 04 18:45:06 crc kubenswrapper[4733]: I1204 18:45:06.358637 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ce93f96-1275-437f-9bcd-a7b299bae008" path="/var/lib/kubelet/pods/4ce93f96-1275-437f-9bcd-a7b299bae008/volumes" Dec 04 18:45:11 crc kubenswrapper[4733]: I1204 18:45:11.336189 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:45:11 crc kubenswrapper[4733]: E1204 18:45:11.337988 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:45:20 crc kubenswrapper[4733]: I1204 18:45:20.326561 4733 scope.go:117] "RemoveContainer" containerID="e5ddd5102295aabdfc76466c7e2b9419618ad91881c05eeb27ba6cac4a5c93ab" Dec 04 18:45:24 crc kubenswrapper[4733]: I1204 18:45:24.335218 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:45:24 crc kubenswrapper[4733]: E1204 18:45:24.335774 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:45:39 crc kubenswrapper[4733]: I1204 18:45:39.336431 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:45:39 crc kubenswrapper[4733]: E1204 18:45:39.337398 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:45:52 crc kubenswrapper[4733]: I1204 18:45:52.336167 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:45:52 crc kubenswrapper[4733]: I1204 18:45:52.756322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"6862ceebd187c03c4a5bbf5aaf8f57024810276fc9d166786d2877e46fe29049"} Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.908843 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kc4ks"] Dec 04 18:47:49 crc kubenswrapper[4733]: E1204 18:47:49.909946 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerName="extract-content" Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.909969 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerName="extract-content" Dec 04 18:47:49 crc kubenswrapper[4733]: E1204 18:47:49.909993 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerName="extract-utilities" Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.910029 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerName="extract-utilities" Dec 04 18:47:49 crc kubenswrapper[4733]: E1204 18:47:49.910061 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfcbd43-0013-4715-81ad-dd378598b7e1" containerName="collect-profiles" Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.910071 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfcbd43-0013-4715-81ad-dd378598b7e1" containerName="collect-profiles" Dec 04 18:47:49 crc kubenswrapper[4733]: E1204 18:47:49.910096 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerName="registry-server" Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.910108 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerName="registry-server" Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.910350 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bfcbd43-0013-4715-81ad-dd378598b7e1" containerName="collect-profiles" Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.910384 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7746e49f-8f15-4abd-938c-2aa5e2bc64fa" containerName="registry-server" Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.912100 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:49 crc kubenswrapper[4733]: I1204 18:47:49.932053 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kc4ks"] Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.042327 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-catalog-content\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.042638 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqskr\" (UniqueName: \"kubernetes.io/projected/8f17a6ad-68ec-4909-b435-1c84cb0d0919-kube-api-access-gqskr\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.042963 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-utilities\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.144831 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-catalog-content\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.144889 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqskr\" (UniqueName: \"kubernetes.io/projected/8f17a6ad-68ec-4909-b435-1c84cb0d0919-kube-api-access-gqskr\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.144943 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-utilities\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.145467 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-utilities\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.145761 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-catalog-content\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.170844 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqskr\" (UniqueName: \"kubernetes.io/projected/8f17a6ad-68ec-4909-b435-1c84cb0d0919-kube-api-access-gqskr\") pod \"redhat-operators-kc4ks\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.246850 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.706627 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kc4ks"] Dec 04 18:47:50 crc kubenswrapper[4733]: I1204 18:47:50.789570 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kc4ks" event={"ID":"8f17a6ad-68ec-4909-b435-1c84cb0d0919","Type":"ContainerStarted","Data":"990696adc6c8350b658a076e1aaf66294b3ca94940c5846ea19291d218974f03"} Dec 04 18:47:51 crc kubenswrapper[4733]: I1204 18:47:51.799541 4733 generic.go:334] "Generic (PLEG): container finished" podID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerID="f89d0d10383234ec3aeb6977c977e3c0247e1aad02e06080c91f0e73c772e692" exitCode=0 Dec 04 18:47:51 crc kubenswrapper[4733]: I1204 18:47:51.799581 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kc4ks" event={"ID":"8f17a6ad-68ec-4909-b435-1c84cb0d0919","Type":"ContainerDied","Data":"f89d0d10383234ec3aeb6977c977e3c0247e1aad02e06080c91f0e73c772e692"} Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.275732 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tqx8c"] Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.277500 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.305742 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tqx8c"] Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.376927 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-utilities\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.377296 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-catalog-content\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.377325 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6sr\" (UniqueName: \"kubernetes.io/projected/2a0c2478-1579-4c28-893c-73ec3d94fbf2-kube-api-access-9n6sr\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.478762 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-catalog-content\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.478821 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6sr\" (UniqueName: \"kubernetes.io/projected/2a0c2478-1579-4c28-893c-73ec3d94fbf2-kube-api-access-9n6sr\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.478888 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-utilities\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.479327 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-catalog-content\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.479376 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-utilities\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.643639 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6sr\" (UniqueName: \"kubernetes.io/projected/2a0c2478-1579-4c28-893c-73ec3d94fbf2-kube-api-access-9n6sr\") pod \"certified-operators-tqx8c\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:52 crc kubenswrapper[4733]: I1204 18:47:52.905834 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:47:53 crc kubenswrapper[4733]: I1204 18:47:53.405187 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tqx8c"] Dec 04 18:47:53 crc kubenswrapper[4733]: W1204 18:47:53.413072 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a0c2478_1579_4c28_893c_73ec3d94fbf2.slice/crio-1ae34caa68b2bd4e9e5dd9337c0cb66116ed7bec11bd368cfd7a628246da91cd WatchSource:0}: Error finding container 1ae34caa68b2bd4e9e5dd9337c0cb66116ed7bec11bd368cfd7a628246da91cd: Status 404 returned error can't find the container with id 1ae34caa68b2bd4e9e5dd9337c0cb66116ed7bec11bd368cfd7a628246da91cd Dec 04 18:47:53 crc kubenswrapper[4733]: I1204 18:47:53.823064 4733 generic.go:334] "Generic (PLEG): container finished" podID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerID="7dfd4c4a5e5abf4164f3a0537f120992195d870be4d39e43dc108b894c8d4a88" exitCode=0 Dec 04 18:47:53 crc kubenswrapper[4733]: I1204 18:47:53.824592 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kc4ks" event={"ID":"8f17a6ad-68ec-4909-b435-1c84cb0d0919","Type":"ContainerDied","Data":"7dfd4c4a5e5abf4164f3a0537f120992195d870be4d39e43dc108b894c8d4a88"} Dec 04 18:47:53 crc kubenswrapper[4733]: I1204 18:47:53.825711 4733 generic.go:334] "Generic (PLEG): container finished" podID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerID="b2c5d0f9b19e57e65262914fcef0b039733108942426fc555f46e9d7941c1171" exitCode=0 Dec 04 18:47:53 crc kubenswrapper[4733]: I1204 18:47:53.825740 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqx8c" event={"ID":"2a0c2478-1579-4c28-893c-73ec3d94fbf2","Type":"ContainerDied","Data":"b2c5d0f9b19e57e65262914fcef0b039733108942426fc555f46e9d7941c1171"} Dec 04 18:47:53 crc kubenswrapper[4733]: I1204 18:47:53.825759 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqx8c" event={"ID":"2a0c2478-1579-4c28-893c-73ec3d94fbf2","Type":"ContainerStarted","Data":"1ae34caa68b2bd4e9e5dd9337c0cb66116ed7bec11bd368cfd7a628246da91cd"} Dec 04 18:47:54 crc kubenswrapper[4733]: I1204 18:47:54.834384 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kc4ks" event={"ID":"8f17a6ad-68ec-4909-b435-1c84cb0d0919","Type":"ContainerStarted","Data":"c03004ad894f39181441ee2a17c12deee03628c40b7378b4135103183eeea0e3"} Dec 04 18:47:54 crc kubenswrapper[4733]: I1204 18:47:54.836359 4733 generic.go:334] "Generic (PLEG): container finished" podID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerID="34be2964198d843d7f6c3dce0a30cf6cdd62283f7a15ef0439546681ee640736" exitCode=0 Dec 04 18:47:54 crc kubenswrapper[4733]: I1204 18:47:54.836394 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqx8c" event={"ID":"2a0c2478-1579-4c28-893c-73ec3d94fbf2","Type":"ContainerDied","Data":"34be2964198d843d7f6c3dce0a30cf6cdd62283f7a15ef0439546681ee640736"} Dec 04 18:47:54 crc kubenswrapper[4733]: I1204 18:47:54.865759 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kc4ks" podStartSLOduration=3.470332005 podStartE2EDuration="5.865739822s" podCreationTimestamp="2025-12-04 18:47:49 +0000 UTC" firstStartedPulling="2025-12-04 18:47:51.801464837 +0000 UTC m=+4133.756825883" lastFinishedPulling="2025-12-04 18:47:54.196872654 +0000 UTC m=+4136.152233700" observedRunningTime="2025-12-04 18:47:54.860017558 +0000 UTC m=+4136.815378654" watchObservedRunningTime="2025-12-04 18:47:54.865739822 +0000 UTC m=+4136.821100868" Dec 04 18:47:55 crc kubenswrapper[4733]: I1204 18:47:55.846394 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqx8c" event={"ID":"2a0c2478-1579-4c28-893c-73ec3d94fbf2","Type":"ContainerStarted","Data":"dc35339d21ac281f4e3eea64688bd85a9457956329e70772d6bc23bf36476558"} Dec 04 18:47:55 crc kubenswrapper[4733]: I1204 18:47:55.863720 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tqx8c" podStartSLOduration=2.467188844 podStartE2EDuration="3.863701799s" podCreationTimestamp="2025-12-04 18:47:52 +0000 UTC" firstStartedPulling="2025-12-04 18:47:53.827021813 +0000 UTC m=+4135.782382869" lastFinishedPulling="2025-12-04 18:47:55.223534778 +0000 UTC m=+4137.178895824" observedRunningTime="2025-12-04 18:47:55.862637351 +0000 UTC m=+4137.817998417" watchObservedRunningTime="2025-12-04 18:47:55.863701799 +0000 UTC m=+4137.819062845" Dec 04 18:48:00 crc kubenswrapper[4733]: I1204 18:48:00.247685 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:48:00 crc kubenswrapper[4733]: I1204 18:48:00.248267 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:48:00 crc kubenswrapper[4733]: I1204 18:48:00.320438 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:48:00 crc kubenswrapper[4733]: I1204 18:48:00.954005 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:48:01 crc kubenswrapper[4733]: I1204 18:48:01.028999 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kc4ks"] Dec 04 18:48:02 crc kubenswrapper[4733]: I1204 18:48:02.905967 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:48:02 crc kubenswrapper[4733]: I1204 18:48:02.906017 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:48:02 crc kubenswrapper[4733]: I1204 18:48:02.906409 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kc4ks" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerName="registry-server" containerID="cri-o://c03004ad894f39181441ee2a17c12deee03628c40b7378b4135103183eeea0e3" gracePeriod=2 Dec 04 18:48:02 crc kubenswrapper[4733]: I1204 18:48:02.947389 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:48:03 crc kubenswrapper[4733]: I1204 18:48:03.914734 4733 generic.go:334] "Generic (PLEG): container finished" podID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerID="c03004ad894f39181441ee2a17c12deee03628c40b7378b4135103183eeea0e3" exitCode=0 Dec 04 18:48:03 crc kubenswrapper[4733]: I1204 18:48:03.914812 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kc4ks" event={"ID":"8f17a6ad-68ec-4909-b435-1c84cb0d0919","Type":"ContainerDied","Data":"c03004ad894f39181441ee2a17c12deee03628c40b7378b4135103183eeea0e3"} Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.401334 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.453025 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tqx8c"] Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.520337 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.662330 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-catalog-content\") pod \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.662389 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-utilities\") pod \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.662542 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqskr\" (UniqueName: \"kubernetes.io/projected/8f17a6ad-68ec-4909-b435-1c84cb0d0919-kube-api-access-gqskr\") pod \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\" (UID: \"8f17a6ad-68ec-4909-b435-1c84cb0d0919\") " Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.663688 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-utilities" (OuterVolumeSpecName: "utilities") pod "8f17a6ad-68ec-4909-b435-1c84cb0d0919" (UID: "8f17a6ad-68ec-4909-b435-1c84cb0d0919"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.664056 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.668657 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f17a6ad-68ec-4909-b435-1c84cb0d0919-kube-api-access-gqskr" (OuterVolumeSpecName: "kube-api-access-gqskr") pod "8f17a6ad-68ec-4909-b435-1c84cb0d0919" (UID: "8f17a6ad-68ec-4909-b435-1c84cb0d0919"). InnerVolumeSpecName "kube-api-access-gqskr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.765626 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqskr\" (UniqueName: \"kubernetes.io/projected/8f17a6ad-68ec-4909-b435-1c84cb0d0919-kube-api-access-gqskr\") on node \"crc\" DevicePath \"\"" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.794524 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f17a6ad-68ec-4909-b435-1c84cb0d0919" (UID: "8f17a6ad-68ec-4909-b435-1c84cb0d0919"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.866513 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f17a6ad-68ec-4909-b435-1c84cb0d0919-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.926709 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kc4ks" event={"ID":"8f17a6ad-68ec-4909-b435-1c84cb0d0919","Type":"ContainerDied","Data":"990696adc6c8350b658a076e1aaf66294b3ca94940c5846ea19291d218974f03"} Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.926841 4733 scope.go:117] "RemoveContainer" containerID="c03004ad894f39181441ee2a17c12deee03628c40b7378b4135103183eeea0e3" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.927101 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kc4ks" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.956147 4733 scope.go:117] "RemoveContainer" containerID="7dfd4c4a5e5abf4164f3a0537f120992195d870be4d39e43dc108b894c8d4a88" Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.968955 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kc4ks"] Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.974841 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kc4ks"] Dec 04 18:48:04 crc kubenswrapper[4733]: I1204 18:48:04.999140 4733 scope.go:117] "RemoveContainer" containerID="f89d0d10383234ec3aeb6977c977e3c0247e1aad02e06080c91f0e73c772e692" Dec 04 18:48:05 crc kubenswrapper[4733]: I1204 18:48:05.939562 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tqx8c" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerName="registry-server" containerID="cri-o://dc35339d21ac281f4e3eea64688bd85a9457956329e70772d6bc23bf36476558" gracePeriod=2 Dec 04 18:48:06 crc kubenswrapper[4733]: I1204 18:48:06.344345 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" path="/var/lib/kubelet/pods/8f17a6ad-68ec-4909-b435-1c84cb0d0919/volumes" Dec 04 18:48:06 crc kubenswrapper[4733]: I1204 18:48:06.950169 4733 generic.go:334] "Generic (PLEG): container finished" podID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerID="dc35339d21ac281f4e3eea64688bd85a9457956329e70772d6bc23bf36476558" exitCode=0 Dec 04 18:48:06 crc kubenswrapper[4733]: I1204 18:48:06.950267 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqx8c" event={"ID":"2a0c2478-1579-4c28-893c-73ec3d94fbf2","Type":"ContainerDied","Data":"dc35339d21ac281f4e3eea64688bd85a9457956329e70772d6bc23bf36476558"} Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.469145 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.608014 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-catalog-content\") pod \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.608066 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-utilities\") pod \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.608114 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n6sr\" (UniqueName: \"kubernetes.io/projected/2a0c2478-1579-4c28-893c-73ec3d94fbf2-kube-api-access-9n6sr\") pod \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\" (UID: \"2a0c2478-1579-4c28-893c-73ec3d94fbf2\") " Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.609407 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-utilities" (OuterVolumeSpecName: "utilities") pod "2a0c2478-1579-4c28-893c-73ec3d94fbf2" (UID: "2a0c2478-1579-4c28-893c-73ec3d94fbf2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.616605 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a0c2478-1579-4c28-893c-73ec3d94fbf2-kube-api-access-9n6sr" (OuterVolumeSpecName: "kube-api-access-9n6sr") pod "2a0c2478-1579-4c28-893c-73ec3d94fbf2" (UID: "2a0c2478-1579-4c28-893c-73ec3d94fbf2"). InnerVolumeSpecName "kube-api-access-9n6sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.669182 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a0c2478-1579-4c28-893c-73ec3d94fbf2" (UID: "2a0c2478-1579-4c28-893c-73ec3d94fbf2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.710232 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.710282 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a0c2478-1579-4c28-893c-73ec3d94fbf2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.710296 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n6sr\" (UniqueName: \"kubernetes.io/projected/2a0c2478-1579-4c28-893c-73ec3d94fbf2-kube-api-access-9n6sr\") on node \"crc\" DevicePath \"\"" Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.964780 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqx8c" event={"ID":"2a0c2478-1579-4c28-893c-73ec3d94fbf2","Type":"ContainerDied","Data":"1ae34caa68b2bd4e9e5dd9337c0cb66116ed7bec11bd368cfd7a628246da91cd"} Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.964904 4733 scope.go:117] "RemoveContainer" containerID="dc35339d21ac281f4e3eea64688bd85a9457956329e70772d6bc23bf36476558" Dec 04 18:48:07 crc kubenswrapper[4733]: I1204 18:48:07.965139 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqx8c" Dec 04 18:48:08 crc kubenswrapper[4733]: I1204 18:48:08.022400 4733 scope.go:117] "RemoveContainer" containerID="34be2964198d843d7f6c3dce0a30cf6cdd62283f7a15ef0439546681ee640736" Dec 04 18:48:08 crc kubenswrapper[4733]: I1204 18:48:08.047852 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tqx8c"] Dec 04 18:48:08 crc kubenswrapper[4733]: I1204 18:48:08.057782 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tqx8c"] Dec 04 18:48:08 crc kubenswrapper[4733]: I1204 18:48:08.085970 4733 scope.go:117] "RemoveContainer" containerID="b2c5d0f9b19e57e65262914fcef0b039733108942426fc555f46e9d7941c1171" Dec 04 18:48:08 crc kubenswrapper[4733]: I1204 18:48:08.346020 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" path="/var/lib/kubelet/pods/2a0c2478-1579-4c28-893c-73ec3d94fbf2/volumes" Dec 04 18:48:15 crc kubenswrapper[4733]: I1204 18:48:15.361841 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:48:15 crc kubenswrapper[4733]: I1204 18:48:15.362669 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:48:45 crc kubenswrapper[4733]: I1204 18:48:45.362313 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:48:45 crc kubenswrapper[4733]: I1204 18:48:45.364179 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:49:15 crc kubenswrapper[4733]: I1204 18:49:15.362442 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:49:15 crc kubenswrapper[4733]: I1204 18:49:15.363083 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:49:15 crc kubenswrapper[4733]: I1204 18:49:15.363136 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:49:15 crc kubenswrapper[4733]: I1204 18:49:15.363935 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6862ceebd187c03c4a5bbf5aaf8f57024810276fc9d166786d2877e46fe29049"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:49:15 crc kubenswrapper[4733]: I1204 18:49:15.364011 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://6862ceebd187c03c4a5bbf5aaf8f57024810276fc9d166786d2877e46fe29049" gracePeriod=600 Dec 04 18:49:15 crc kubenswrapper[4733]: I1204 18:49:15.516637 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="6862ceebd187c03c4a5bbf5aaf8f57024810276fc9d166786d2877e46fe29049" exitCode=0 Dec 04 18:49:15 crc kubenswrapper[4733]: I1204 18:49:15.516663 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"6862ceebd187c03c4a5bbf5aaf8f57024810276fc9d166786d2877e46fe29049"} Dec 04 18:49:15 crc kubenswrapper[4733]: I1204 18:49:15.516748 4733 scope.go:117] "RemoveContainer" containerID="fb7c2a569b5e56c658f750ffb24fdcaf0fd489fbcdf93532a0680cd7672b8b71" Dec 04 18:49:16 crc kubenswrapper[4733]: I1204 18:49:16.528775 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a"} Dec 04 18:51:15 crc kubenswrapper[4733]: I1204 18:51:15.361600 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:51:15 crc kubenswrapper[4733]: I1204 18:51:15.362194 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:51:45 crc kubenswrapper[4733]: I1204 18:51:45.361881 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:51:45 crc kubenswrapper[4733]: I1204 18:51:45.362315 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.768957 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4s7rj"] Dec 04 18:52:04 crc kubenswrapper[4733]: E1204 18:52:04.772482 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerName="extract-utilities" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.772661 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerName="extract-utilities" Dec 04 18:52:04 crc kubenswrapper[4733]: E1204 18:52:04.772838 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerName="registry-server" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.772978 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerName="registry-server" Dec 04 18:52:04 crc kubenswrapper[4733]: E1204 18:52:04.773126 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerName="registry-server" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.773253 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerName="registry-server" Dec 04 18:52:04 crc kubenswrapper[4733]: E1204 18:52:04.773413 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerName="extract-utilities" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.773553 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerName="extract-utilities" Dec 04 18:52:04 crc kubenswrapper[4733]: E1204 18:52:04.773696 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerName="extract-content" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.773875 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerName="extract-content" Dec 04 18:52:04 crc kubenswrapper[4733]: E1204 18:52:04.774072 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerName="extract-content" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.774207 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerName="extract-content" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.774611 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a0c2478-1579-4c28-893c-73ec3d94fbf2" containerName="registry-server" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.774833 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f17a6ad-68ec-4909-b435-1c84cb0d0919" containerName="registry-server" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.777311 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s7rj"] Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.777506 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.865629 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpx82\" (UniqueName: \"kubernetes.io/projected/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-kube-api-access-wpx82\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.866027 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-utilities\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.866187 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-catalog-content\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.967383 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-catalog-content\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.967481 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpx82\" (UniqueName: \"kubernetes.io/projected/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-kube-api-access-wpx82\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.967548 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-utilities\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.968030 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-catalog-content\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.968075 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-utilities\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:04 crc kubenswrapper[4733]: I1204 18:52:04.992372 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpx82\" (UniqueName: \"kubernetes.io/projected/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-kube-api-access-wpx82\") pod \"redhat-marketplace-4s7rj\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:05 crc kubenswrapper[4733]: I1204 18:52:05.119416 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:05 crc kubenswrapper[4733]: I1204 18:52:05.531669 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s7rj"] Dec 04 18:52:06 crc kubenswrapper[4733]: I1204 18:52:06.032025 4733 generic.go:334] "Generic (PLEG): container finished" podID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerID="ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372" exitCode=0 Dec 04 18:52:06 crc kubenswrapper[4733]: I1204 18:52:06.032079 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s7rj" event={"ID":"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654","Type":"ContainerDied","Data":"ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372"} Dec 04 18:52:06 crc kubenswrapper[4733]: I1204 18:52:06.033476 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s7rj" event={"ID":"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654","Type":"ContainerStarted","Data":"e1cf51ecd8ccce7951830dbdea5b6e8e938cdd70a24b54c588193ee93ab71460"} Dec 04 18:52:06 crc kubenswrapper[4733]: I1204 18:52:06.034199 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:52:07 crc kubenswrapper[4733]: I1204 18:52:07.042335 4733 generic.go:334] "Generic (PLEG): container finished" podID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerID="f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1" exitCode=0 Dec 04 18:52:07 crc kubenswrapper[4733]: I1204 18:52:07.042398 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s7rj" event={"ID":"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654","Type":"ContainerDied","Data":"f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1"} Dec 04 18:52:08 crc kubenswrapper[4733]: I1204 18:52:08.052140 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s7rj" event={"ID":"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654","Type":"ContainerStarted","Data":"29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225"} Dec 04 18:52:08 crc kubenswrapper[4733]: I1204 18:52:08.076280 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4s7rj" podStartSLOduration=2.414271566 podStartE2EDuration="4.076254529s" podCreationTimestamp="2025-12-04 18:52:04 +0000 UTC" firstStartedPulling="2025-12-04 18:52:06.033884329 +0000 UTC m=+4387.989245395" lastFinishedPulling="2025-12-04 18:52:07.695867312 +0000 UTC m=+4389.651228358" observedRunningTime="2025-12-04 18:52:08.071150911 +0000 UTC m=+4390.026511957" watchObservedRunningTime="2025-12-04 18:52:08.076254529 +0000 UTC m=+4390.031615595" Dec 04 18:52:15 crc kubenswrapper[4733]: I1204 18:52:15.120106 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:15 crc kubenswrapper[4733]: I1204 18:52:15.120730 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:15 crc kubenswrapper[4733]: I1204 18:52:15.170851 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:15 crc kubenswrapper[4733]: I1204 18:52:15.361910 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:52:15 crc kubenswrapper[4733]: I1204 18:52:15.361986 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 18:52:15 crc kubenswrapper[4733]: I1204 18:52:15.362034 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 18:52:15 crc kubenswrapper[4733]: I1204 18:52:15.362628 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 18:52:15 crc kubenswrapper[4733]: I1204 18:52:15.362688 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" gracePeriod=600 Dec 04 18:52:15 crc kubenswrapper[4733]: E1204 18:52:15.995095 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:52:16 crc kubenswrapper[4733]: I1204 18:52:16.124208 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" exitCode=0 Dec 04 18:52:16 crc kubenswrapper[4733]: I1204 18:52:16.124282 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a"} Dec 04 18:52:16 crc kubenswrapper[4733]: I1204 18:52:16.124333 4733 scope.go:117] "RemoveContainer" containerID="6862ceebd187c03c4a5bbf5aaf8f57024810276fc9d166786d2877e46fe29049" Dec 04 18:52:16 crc kubenswrapper[4733]: I1204 18:52:16.125215 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:52:16 crc kubenswrapper[4733]: E1204 18:52:16.125563 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:52:16 crc kubenswrapper[4733]: I1204 18:52:16.185552 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:16 crc kubenswrapper[4733]: I1204 18:52:16.230858 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s7rj"] Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.152739 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4s7rj" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerName="registry-server" containerID="cri-o://29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225" gracePeriod=2 Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.578209 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.683535 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-utilities\") pod \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.683877 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-catalog-content\") pod \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.684048 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpx82\" (UniqueName: \"kubernetes.io/projected/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-kube-api-access-wpx82\") pod \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\" (UID: \"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654\") " Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.684900 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-utilities" (OuterVolumeSpecName: "utilities") pod "8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" (UID: "8d91d8fb-6fd7-4fd8-b4df-e4599dff3654"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.689180 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-kube-api-access-wpx82" (OuterVolumeSpecName: "kube-api-access-wpx82") pod "8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" (UID: "8d91d8fb-6fd7-4fd8-b4df-e4599dff3654"). InnerVolumeSpecName "kube-api-access-wpx82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.706352 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" (UID: "8d91d8fb-6fd7-4fd8-b4df-e4599dff3654"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.784858 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpx82\" (UniqueName: \"kubernetes.io/projected/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-kube-api-access-wpx82\") on node \"crc\" DevicePath \"\"" Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.784893 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:52:18 crc kubenswrapper[4733]: I1204 18:52:18.784905 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.163978 4733 generic.go:334] "Generic (PLEG): container finished" podID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerID="29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225" exitCode=0 Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.164060 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s7rj" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.164062 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s7rj" event={"ID":"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654","Type":"ContainerDied","Data":"29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225"} Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.164262 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s7rj" event={"ID":"8d91d8fb-6fd7-4fd8-b4df-e4599dff3654","Type":"ContainerDied","Data":"e1cf51ecd8ccce7951830dbdea5b6e8e938cdd70a24b54c588193ee93ab71460"} Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.164310 4733 scope.go:117] "RemoveContainer" containerID="29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.193655 4733 scope.go:117] "RemoveContainer" containerID="f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.215590 4733 scope.go:117] "RemoveContainer" containerID="ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.226733 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s7rj"] Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.234774 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s7rj"] Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.256417 4733 scope.go:117] "RemoveContainer" containerID="29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225" Dec 04 18:52:19 crc kubenswrapper[4733]: E1204 18:52:19.257074 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225\": container with ID starting with 29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225 not found: ID does not exist" containerID="29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.257152 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225"} err="failed to get container status \"29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225\": rpc error: code = NotFound desc = could not find container \"29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225\": container with ID starting with 29e9e887dd1175b1bc8abeb6a6ee1a6d159ce0e3ae8801b20d36ca6e040cf225 not found: ID does not exist" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.257219 4733 scope.go:117] "RemoveContainer" containerID="f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1" Dec 04 18:52:19 crc kubenswrapper[4733]: E1204 18:52:19.257937 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1\": container with ID starting with f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1 not found: ID does not exist" containerID="f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.257997 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1"} err="failed to get container status \"f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1\": rpc error: code = NotFound desc = could not find container \"f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1\": container with ID starting with f2f2caba748177fda589ccc8fbd68a68a1de67e7d11c08c602adcfa193bf00d1 not found: ID does not exist" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.258019 4733 scope.go:117] "RemoveContainer" containerID="ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372" Dec 04 18:52:19 crc kubenswrapper[4733]: E1204 18:52:19.258574 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372\": container with ID starting with ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372 not found: ID does not exist" containerID="ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372" Dec 04 18:52:19 crc kubenswrapper[4733]: I1204 18:52:19.258682 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372"} err="failed to get container status \"ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372\": rpc error: code = NotFound desc = could not find container \"ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372\": container with ID starting with ed5c2f8133ac9dfe99f8813ee7cbf49059b06338708ce425cd66a95870e95372 not found: ID does not exist" Dec 04 18:52:20 crc kubenswrapper[4733]: I1204 18:52:20.348140 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" path="/var/lib/kubelet/pods/8d91d8fb-6fd7-4fd8-b4df-e4599dff3654/volumes" Dec 04 18:52:26 crc kubenswrapper[4733]: I1204 18:52:26.335723 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:52:26 crc kubenswrapper[4733]: E1204 18:52:26.336751 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:52:37 crc kubenswrapper[4733]: I1204 18:52:37.335488 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:52:37 crc kubenswrapper[4733]: E1204 18:52:37.336274 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:52:48 crc kubenswrapper[4733]: I1204 18:52:48.343182 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:52:48 crc kubenswrapper[4733]: E1204 18:52:48.344892 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:53:00 crc kubenswrapper[4733]: I1204 18:53:00.335566 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:53:00 crc kubenswrapper[4733]: E1204 18:53:00.337738 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:53:15 crc kubenswrapper[4733]: I1204 18:53:15.335296 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:53:15 crc kubenswrapper[4733]: E1204 18:53:15.336185 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:53:28 crc kubenswrapper[4733]: I1204 18:53:28.339262 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:53:28 crc kubenswrapper[4733]: E1204 18:53:28.340019 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:53:42 crc kubenswrapper[4733]: I1204 18:53:42.335733 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:53:42 crc kubenswrapper[4733]: E1204 18:53:42.336558 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:53:57 crc kubenswrapper[4733]: I1204 18:53:57.335103 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:53:57 crc kubenswrapper[4733]: E1204 18:53:57.335922 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:54:10 crc kubenswrapper[4733]: I1204 18:54:10.336498 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:54:10 crc kubenswrapper[4733]: E1204 18:54:10.337785 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:54:21 crc kubenswrapper[4733]: I1204 18:54:21.335538 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:54:21 crc kubenswrapper[4733]: E1204 18:54:21.336445 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:54:35 crc kubenswrapper[4733]: I1204 18:54:35.335774 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:54:35 crc kubenswrapper[4733]: E1204 18:54:35.336952 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:54:49 crc kubenswrapper[4733]: I1204 18:54:49.336028 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:54:49 crc kubenswrapper[4733]: E1204 18:54:49.337083 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:55:03 crc kubenswrapper[4733]: I1204 18:55:03.336084 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:55:03 crc kubenswrapper[4733]: E1204 18:55:03.337164 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.056697 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qxzmq"] Dec 04 18:55:07 crc kubenswrapper[4733]: E1204 18:55:07.061200 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerName="registry-server" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.061251 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerName="registry-server" Dec 04 18:55:07 crc kubenswrapper[4733]: E1204 18:55:07.061281 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerName="extract-content" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.061299 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerName="extract-content" Dec 04 18:55:07 crc kubenswrapper[4733]: E1204 18:55:07.061356 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerName="extract-utilities" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.061376 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerName="extract-utilities" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.061822 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d91d8fb-6fd7-4fd8-b4df-e4599dff3654" containerName="registry-server" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.064359 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.066883 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qxzmq"] Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.210960 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-utilities\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.211034 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qjs2\" (UniqueName: \"kubernetes.io/projected/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-kube-api-access-9qjs2\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.211102 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-catalog-content\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.312762 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qjs2\" (UniqueName: \"kubernetes.io/projected/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-kube-api-access-9qjs2\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.312868 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-catalog-content\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.312936 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-utilities\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.313555 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-utilities\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.313559 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-catalog-content\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.341050 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qjs2\" (UniqueName: \"kubernetes.io/projected/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-kube-api-access-9qjs2\") pod \"community-operators-qxzmq\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.418320 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:07 crc kubenswrapper[4733]: I1204 18:55:07.944208 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qxzmq"] Dec 04 18:55:08 crc kubenswrapper[4733]: I1204 18:55:08.599093 4733 generic.go:334] "Generic (PLEG): container finished" podID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerID="9075c46c1fe5f6f3c365ff35058e62c0b05cb829ac674826b5bffd34d25b2179" exitCode=0 Dec 04 18:55:08 crc kubenswrapper[4733]: I1204 18:55:08.599172 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxzmq" event={"ID":"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25","Type":"ContainerDied","Data":"9075c46c1fe5f6f3c365ff35058e62c0b05cb829ac674826b5bffd34d25b2179"} Dec 04 18:55:08 crc kubenswrapper[4733]: I1204 18:55:08.599429 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxzmq" event={"ID":"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25","Type":"ContainerStarted","Data":"c3bb08ddb5550bac8ff0f66d429cc56bb6bf7a5423da22f9dd43aae1a740328e"} Dec 04 18:55:09 crc kubenswrapper[4733]: I1204 18:55:09.610046 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxzmq" event={"ID":"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25","Type":"ContainerStarted","Data":"dc0c0fd12806c7d5f4114a9d8041793be077efad8d8e0cc444d78c980652810b"} Dec 04 18:55:10 crc kubenswrapper[4733]: I1204 18:55:10.626112 4733 generic.go:334] "Generic (PLEG): container finished" podID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerID="dc0c0fd12806c7d5f4114a9d8041793be077efad8d8e0cc444d78c980652810b" exitCode=0 Dec 04 18:55:10 crc kubenswrapper[4733]: I1204 18:55:10.626196 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxzmq" event={"ID":"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25","Type":"ContainerDied","Data":"dc0c0fd12806c7d5f4114a9d8041793be077efad8d8e0cc444d78c980652810b"} Dec 04 18:55:11 crc kubenswrapper[4733]: I1204 18:55:11.637378 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxzmq" event={"ID":"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25","Type":"ContainerStarted","Data":"dfa870050042792514c81263bc9e8edac4acb86800b34461255a3e411b61105f"} Dec 04 18:55:11 crc kubenswrapper[4733]: I1204 18:55:11.653504 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qxzmq" podStartSLOduration=2.214812654 podStartE2EDuration="4.653473893s" podCreationTimestamp="2025-12-04 18:55:07 +0000 UTC" firstStartedPulling="2025-12-04 18:55:08.601224245 +0000 UTC m=+4570.556585311" lastFinishedPulling="2025-12-04 18:55:11.039885474 +0000 UTC m=+4572.995246550" observedRunningTime="2025-12-04 18:55:11.652836725 +0000 UTC m=+4573.608197791" watchObservedRunningTime="2025-12-04 18:55:11.653473893 +0000 UTC m=+4573.608834949" Dec 04 18:55:17 crc kubenswrapper[4733]: I1204 18:55:17.419029 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:17 crc kubenswrapper[4733]: I1204 18:55:17.419791 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:17 crc kubenswrapper[4733]: I1204 18:55:17.463739 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:17 crc kubenswrapper[4733]: I1204 18:55:17.736027 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:18 crc kubenswrapper[4733]: I1204 18:55:18.338553 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:55:18 crc kubenswrapper[4733]: E1204 18:55:18.338769 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:55:22 crc kubenswrapper[4733]: I1204 18:55:22.222734 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qxzmq"] Dec 04 18:55:22 crc kubenswrapper[4733]: I1204 18:55:22.223225 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qxzmq" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerName="registry-server" containerID="cri-o://dfa870050042792514c81263bc9e8edac4acb86800b34461255a3e411b61105f" gracePeriod=2 Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.738242 4733 generic.go:334] "Generic (PLEG): container finished" podID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerID="dfa870050042792514c81263bc9e8edac4acb86800b34461255a3e411b61105f" exitCode=0 Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.738293 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxzmq" event={"ID":"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25","Type":"ContainerDied","Data":"dfa870050042792514c81263bc9e8edac4acb86800b34461255a3e411b61105f"} Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.794842 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.887131 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-catalog-content\") pod \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.887182 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-utilities\") pod \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.887291 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qjs2\" (UniqueName: \"kubernetes.io/projected/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-kube-api-access-9qjs2\") pod \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\" (UID: \"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25\") " Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.888511 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-utilities" (OuterVolumeSpecName: "utilities") pod "ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" (UID: "ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.895946 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-kube-api-access-9qjs2" (OuterVolumeSpecName: "kube-api-access-9qjs2") pod "ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" (UID: "ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25"). InnerVolumeSpecName "kube-api-access-9qjs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.940462 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" (UID: "ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.989731 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.989823 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:55:23 crc kubenswrapper[4733]: I1204 18:55:23.989841 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qjs2\" (UniqueName: \"kubernetes.io/projected/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25-kube-api-access-9qjs2\") on node \"crc\" DevicePath \"\"" Dec 04 18:55:24 crc kubenswrapper[4733]: I1204 18:55:24.750217 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxzmq" event={"ID":"ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25","Type":"ContainerDied","Data":"c3bb08ddb5550bac8ff0f66d429cc56bb6bf7a5423da22f9dd43aae1a740328e"} Dec 04 18:55:24 crc kubenswrapper[4733]: I1204 18:55:24.750713 4733 scope.go:117] "RemoveContainer" containerID="dfa870050042792514c81263bc9e8edac4acb86800b34461255a3e411b61105f" Dec 04 18:55:24 crc kubenswrapper[4733]: I1204 18:55:24.750312 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxzmq" Dec 04 18:55:24 crc kubenswrapper[4733]: I1204 18:55:24.783702 4733 scope.go:117] "RemoveContainer" containerID="dc0c0fd12806c7d5f4114a9d8041793be077efad8d8e0cc444d78c980652810b" Dec 04 18:55:24 crc kubenswrapper[4733]: I1204 18:55:24.786069 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qxzmq"] Dec 04 18:55:24 crc kubenswrapper[4733]: I1204 18:55:24.813757 4733 scope.go:117] "RemoveContainer" containerID="9075c46c1fe5f6f3c365ff35058e62c0b05cb829ac674826b5bffd34d25b2179" Dec 04 18:55:24 crc kubenswrapper[4733]: I1204 18:55:24.818444 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qxzmq"] Dec 04 18:55:26 crc kubenswrapper[4733]: I1204 18:55:26.356533 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" path="/var/lib/kubelet/pods/ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25/volumes" Dec 04 18:55:32 crc kubenswrapper[4733]: I1204 18:55:32.335568 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:55:32 crc kubenswrapper[4733]: E1204 18:55:32.336594 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:55:44 crc kubenswrapper[4733]: I1204 18:55:44.337205 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:55:44 crc kubenswrapper[4733]: E1204 18:55:44.338108 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:55:56 crc kubenswrapper[4733]: I1204 18:55:56.336096 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:55:56 crc kubenswrapper[4733]: E1204 18:55:56.337165 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:56:11 crc kubenswrapper[4733]: I1204 18:56:11.336023 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:56:11 crc kubenswrapper[4733]: E1204 18:56:11.336836 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:56:25 crc kubenswrapper[4733]: I1204 18:56:25.335529 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:56:25 crc kubenswrapper[4733]: E1204 18:56:25.336737 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:56:39 crc kubenswrapper[4733]: I1204 18:56:39.336489 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:56:39 crc kubenswrapper[4733]: E1204 18:56:39.337553 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:56:50 crc kubenswrapper[4733]: I1204 18:56:50.336113 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:56:50 crc kubenswrapper[4733]: E1204 18:56:50.337371 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:57:05 crc kubenswrapper[4733]: I1204 18:57:05.335350 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:57:05 crc kubenswrapper[4733]: E1204 18:57:05.336396 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 18:57:19 crc kubenswrapper[4733]: I1204 18:57:19.336055 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 18:57:20 crc kubenswrapper[4733]: I1204 18:57:20.731311 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"133576a4fa20da80bc0ad5a9ec576a91784d7d1b185b32466a081bc4c9e98f7f"} Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.185243 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-t5jhj"] Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.198530 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-t5jhj"] Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.321874 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-tkljl"] Dec 04 18:57:34 crc kubenswrapper[4733]: E1204 18:57:34.322206 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerName="registry-server" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.322229 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerName="registry-server" Dec 04 18:57:34 crc kubenswrapper[4733]: E1204 18:57:34.322245 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerName="extract-utilities" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.322251 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerName="extract-utilities" Dec 04 18:57:34 crc kubenswrapper[4733]: E1204 18:57:34.322267 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerName="extract-content" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.322273 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerName="extract-content" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.322424 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff96bb0c-ea7f-45f0-bd60-5c4f04a64f25" containerName="registry-server" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.323298 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.326599 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.326732 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.326963 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.326746 4733 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-r9cpb" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.359143 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5addc135-8f0f-4edd-ad29-76f3fd340827" path="/var/lib/kubelet/pods/5addc135-8f0f-4edd-ad29-76f3fd340827/volumes" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.360200 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-tkljl"] Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.443390 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/994e15d1-8444-4a0e-b216-3e4d75dea518-crc-storage\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.443500 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/994e15d1-8444-4a0e-b216-3e4d75dea518-node-mnt\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.443539 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhhp7\" (UniqueName: \"kubernetes.io/projected/994e15d1-8444-4a0e-b216-3e4d75dea518-kube-api-access-hhhp7\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.545168 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/994e15d1-8444-4a0e-b216-3e4d75dea518-crc-storage\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.545476 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/994e15d1-8444-4a0e-b216-3e4d75dea518-node-mnt\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.545573 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhhp7\" (UniqueName: \"kubernetes.io/projected/994e15d1-8444-4a0e-b216-3e4d75dea518-kube-api-access-hhhp7\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.546067 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/994e15d1-8444-4a0e-b216-3e4d75dea518-crc-storage\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.546243 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/994e15d1-8444-4a0e-b216-3e4d75dea518-node-mnt\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.564393 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhhp7\" (UniqueName: \"kubernetes.io/projected/994e15d1-8444-4a0e-b216-3e4d75dea518-kube-api-access-hhhp7\") pod \"crc-storage-crc-tkljl\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:34 crc kubenswrapper[4733]: I1204 18:57:34.641673 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:35 crc kubenswrapper[4733]: I1204 18:57:35.119307 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-tkljl"] Dec 04 18:57:35 crc kubenswrapper[4733]: W1204 18:57:35.127379 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod994e15d1_8444_4a0e_b216_3e4d75dea518.slice/crio-36337de88e132a85fca9a97e5379fae21f14b41e3b9a3f94bf284fed7e377130 WatchSource:0}: Error finding container 36337de88e132a85fca9a97e5379fae21f14b41e3b9a3f94bf284fed7e377130: Status 404 returned error can't find the container with id 36337de88e132a85fca9a97e5379fae21f14b41e3b9a3f94bf284fed7e377130 Dec 04 18:57:35 crc kubenswrapper[4733]: I1204 18:57:35.132324 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 18:57:35 crc kubenswrapper[4733]: I1204 18:57:35.853439 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tkljl" event={"ID":"994e15d1-8444-4a0e-b216-3e4d75dea518","Type":"ContainerStarted","Data":"66241135ddd16a802ad2cb212387ee2705c41a0f0015af7e04ba7cdfd47717cb"} Dec 04 18:57:35 crc kubenswrapper[4733]: I1204 18:57:35.853986 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tkljl" event={"ID":"994e15d1-8444-4a0e-b216-3e4d75dea518","Type":"ContainerStarted","Data":"36337de88e132a85fca9a97e5379fae21f14b41e3b9a3f94bf284fed7e377130"} Dec 04 18:57:35 crc kubenswrapper[4733]: I1204 18:57:35.881985 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-tkljl" podStartSLOduration=1.379509338 podStartE2EDuration="1.881954467s" podCreationTimestamp="2025-12-04 18:57:34 +0000 UTC" firstStartedPulling="2025-12-04 18:57:35.131659741 +0000 UTC m=+4717.087020827" lastFinishedPulling="2025-12-04 18:57:35.63410489 +0000 UTC m=+4717.589465956" observedRunningTime="2025-12-04 18:57:35.869646165 +0000 UTC m=+4717.825007221" watchObservedRunningTime="2025-12-04 18:57:35.881954467 +0000 UTC m=+4717.837315533" Dec 04 18:57:36 crc kubenswrapper[4733]: I1204 18:57:36.863675 4733 generic.go:334] "Generic (PLEG): container finished" podID="994e15d1-8444-4a0e-b216-3e4d75dea518" containerID="66241135ddd16a802ad2cb212387ee2705c41a0f0015af7e04ba7cdfd47717cb" exitCode=0 Dec 04 18:57:36 crc kubenswrapper[4733]: I1204 18:57:36.863725 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tkljl" event={"ID":"994e15d1-8444-4a0e-b216-3e4d75dea518","Type":"ContainerDied","Data":"66241135ddd16a802ad2cb212387ee2705c41a0f0015af7e04ba7cdfd47717cb"} Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.170122 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.213233 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/994e15d1-8444-4a0e-b216-3e4d75dea518-crc-storage\") pod \"994e15d1-8444-4a0e-b216-3e4d75dea518\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.213314 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhhp7\" (UniqueName: \"kubernetes.io/projected/994e15d1-8444-4a0e-b216-3e4d75dea518-kube-api-access-hhhp7\") pod \"994e15d1-8444-4a0e-b216-3e4d75dea518\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.213336 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/994e15d1-8444-4a0e-b216-3e4d75dea518-node-mnt\") pod \"994e15d1-8444-4a0e-b216-3e4d75dea518\" (UID: \"994e15d1-8444-4a0e-b216-3e4d75dea518\") " Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.213709 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/994e15d1-8444-4a0e-b216-3e4d75dea518-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "994e15d1-8444-4a0e-b216-3e4d75dea518" (UID: "994e15d1-8444-4a0e-b216-3e4d75dea518"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.225976 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/994e15d1-8444-4a0e-b216-3e4d75dea518-kube-api-access-hhhp7" (OuterVolumeSpecName: "kube-api-access-hhhp7") pod "994e15d1-8444-4a0e-b216-3e4d75dea518" (UID: "994e15d1-8444-4a0e-b216-3e4d75dea518"). InnerVolumeSpecName "kube-api-access-hhhp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.287553 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/994e15d1-8444-4a0e-b216-3e4d75dea518-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "994e15d1-8444-4a0e-b216-3e4d75dea518" (UID: "994e15d1-8444-4a0e-b216-3e4d75dea518"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.317570 4733 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/994e15d1-8444-4a0e-b216-3e4d75dea518-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.317604 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhhp7\" (UniqueName: \"kubernetes.io/projected/994e15d1-8444-4a0e-b216-3e4d75dea518-kube-api-access-hhhp7\") on node \"crc\" DevicePath \"\"" Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.317614 4733 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/994e15d1-8444-4a0e-b216-3e4d75dea518-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.880254 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tkljl" event={"ID":"994e15d1-8444-4a0e-b216-3e4d75dea518","Type":"ContainerDied","Data":"36337de88e132a85fca9a97e5379fae21f14b41e3b9a3f94bf284fed7e377130"} Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.880606 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36337de88e132a85fca9a97e5379fae21f14b41e3b9a3f94bf284fed7e377130" Dec 04 18:57:38 crc kubenswrapper[4733]: I1204 18:57:38.880338 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tkljl" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.153994 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-tkljl"] Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.160215 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-tkljl"] Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.297553 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-tkjkm"] Dec 04 18:57:40 crc kubenswrapper[4733]: E1204 18:57:40.297881 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994e15d1-8444-4a0e-b216-3e4d75dea518" containerName="storage" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.297897 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="994e15d1-8444-4a0e-b216-3e4d75dea518" containerName="storage" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.298067 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="994e15d1-8444-4a0e-b216-3e4d75dea518" containerName="storage" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.298591 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.301259 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.301640 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.301946 4733 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-r9cpb" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.302195 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.319459 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-tkjkm"] Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.353620 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqccf\" (UniqueName: \"kubernetes.io/projected/29b8799f-13a5-4222-b933-be9f2091a616-kube-api-access-sqccf\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.353836 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/29b8799f-13a5-4222-b933-be9f2091a616-node-mnt\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.353869 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/29b8799f-13a5-4222-b933-be9f2091a616-crc-storage\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.361016 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="994e15d1-8444-4a0e-b216-3e4d75dea518" path="/var/lib/kubelet/pods/994e15d1-8444-4a0e-b216-3e4d75dea518/volumes" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.455075 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqccf\" (UniqueName: \"kubernetes.io/projected/29b8799f-13a5-4222-b933-be9f2091a616-kube-api-access-sqccf\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.455173 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/29b8799f-13a5-4222-b933-be9f2091a616-node-mnt\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.455193 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/29b8799f-13a5-4222-b933-be9f2091a616-crc-storage\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.456145 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/29b8799f-13a5-4222-b933-be9f2091a616-node-mnt\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.456870 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/29b8799f-13a5-4222-b933-be9f2091a616-crc-storage\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.474240 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqccf\" (UniqueName: \"kubernetes.io/projected/29b8799f-13a5-4222-b933-be9f2091a616-kube-api-access-sqccf\") pod \"crc-storage-crc-tkjkm\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:40 crc kubenswrapper[4733]: I1204 18:57:40.615722 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:41 crc kubenswrapper[4733]: I1204 18:57:41.034761 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-tkjkm"] Dec 04 18:57:41 crc kubenswrapper[4733]: I1204 18:57:41.910124 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tkjkm" event={"ID":"29b8799f-13a5-4222-b933-be9f2091a616","Type":"ContainerStarted","Data":"3f601e3111f0610c1adfa5e966898af2a093ae136496f434abaa1f6b39970e49"} Dec 04 18:57:42 crc kubenswrapper[4733]: I1204 18:57:42.918455 4733 generic.go:334] "Generic (PLEG): container finished" podID="29b8799f-13a5-4222-b933-be9f2091a616" containerID="f59bb3f79375ef94c299d99a454c10912a9518f555878fc94455460e71e98a7e" exitCode=0 Dec 04 18:57:42 crc kubenswrapper[4733]: I1204 18:57:42.918544 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tkjkm" event={"ID":"29b8799f-13a5-4222-b933-be9f2091a616","Type":"ContainerDied","Data":"f59bb3f79375ef94c299d99a454c10912a9518f555878fc94455460e71e98a7e"} Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.241168 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.314019 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/29b8799f-13a5-4222-b933-be9f2091a616-node-mnt\") pod \"29b8799f-13a5-4222-b933-be9f2091a616\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.314133 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29b8799f-13a5-4222-b933-be9f2091a616-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "29b8799f-13a5-4222-b933-be9f2091a616" (UID: "29b8799f-13a5-4222-b933-be9f2091a616"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.314175 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqccf\" (UniqueName: \"kubernetes.io/projected/29b8799f-13a5-4222-b933-be9f2091a616-kube-api-access-sqccf\") pod \"29b8799f-13a5-4222-b933-be9f2091a616\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.314230 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/29b8799f-13a5-4222-b933-be9f2091a616-crc-storage\") pod \"29b8799f-13a5-4222-b933-be9f2091a616\" (UID: \"29b8799f-13a5-4222-b933-be9f2091a616\") " Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.314604 4733 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/29b8799f-13a5-4222-b933-be9f2091a616-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.320878 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29b8799f-13a5-4222-b933-be9f2091a616-kube-api-access-sqccf" (OuterVolumeSpecName: "kube-api-access-sqccf") pod "29b8799f-13a5-4222-b933-be9f2091a616" (UID: "29b8799f-13a5-4222-b933-be9f2091a616"). InnerVolumeSpecName "kube-api-access-sqccf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.337018 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29b8799f-13a5-4222-b933-be9f2091a616-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "29b8799f-13a5-4222-b933-be9f2091a616" (UID: "29b8799f-13a5-4222-b933-be9f2091a616"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.416694 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqccf\" (UniqueName: \"kubernetes.io/projected/29b8799f-13a5-4222-b933-be9f2091a616-kube-api-access-sqccf\") on node \"crc\" DevicePath \"\"" Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.416732 4733 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/29b8799f-13a5-4222-b933-be9f2091a616-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.935502 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tkjkm" event={"ID":"29b8799f-13a5-4222-b933-be9f2091a616","Type":"ContainerDied","Data":"3f601e3111f0610c1adfa5e966898af2a093ae136496f434abaa1f6b39970e49"} Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.935554 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f601e3111f0610c1adfa5e966898af2a093ae136496f434abaa1f6b39970e49" Dec 04 18:57:44 crc kubenswrapper[4733]: I1204 18:57:44.935555 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tkjkm" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.200678 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pkth4"] Dec 04 18:58:15 crc kubenswrapper[4733]: E1204 18:58:15.202086 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b8799f-13a5-4222-b933-be9f2091a616" containerName="storage" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.202121 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b8799f-13a5-4222-b933-be9f2091a616" containerName="storage" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.202541 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="29b8799f-13a5-4222-b933-be9f2091a616" containerName="storage" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.205035 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.223043 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pkth4"] Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.345646 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgtk9\" (UniqueName: \"kubernetes.io/projected/b8777330-05db-4780-9e3d-2f76a0b7a590-kube-api-access-lgtk9\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.345776 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-utilities\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.345950 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-catalog-content\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.447004 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-utilities\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.447069 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-catalog-content\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.447137 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgtk9\" (UniqueName: \"kubernetes.io/projected/b8777330-05db-4780-9e3d-2f76a0b7a590-kube-api-access-lgtk9\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.447693 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-utilities\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.447755 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-catalog-content\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.474837 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgtk9\" (UniqueName: \"kubernetes.io/projected/b8777330-05db-4780-9e3d-2f76a0b7a590-kube-api-access-lgtk9\") pod \"redhat-operators-pkth4\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.545487 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:15 crc kubenswrapper[4733]: I1204 18:58:15.985687 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pkth4"] Dec 04 18:58:16 crc kubenswrapper[4733]: I1204 18:58:16.220511 4733 generic.go:334] "Generic (PLEG): container finished" podID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerID="b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52" exitCode=0 Dec 04 18:58:16 crc kubenswrapper[4733]: I1204 18:58:16.220681 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pkth4" event={"ID":"b8777330-05db-4780-9e3d-2f76a0b7a590","Type":"ContainerDied","Data":"b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52"} Dec 04 18:58:16 crc kubenswrapper[4733]: I1204 18:58:16.220910 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pkth4" event={"ID":"b8777330-05db-4780-9e3d-2f76a0b7a590","Type":"ContainerStarted","Data":"69212896ba936a383614b842eb67f25d2db4ead12e49ddf5f4a0d8e7b52709d6"} Dec 04 18:58:17 crc kubenswrapper[4733]: I1204 18:58:17.254579 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pkth4" event={"ID":"b8777330-05db-4780-9e3d-2f76a0b7a590","Type":"ContainerStarted","Data":"4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021"} Dec 04 18:58:18 crc kubenswrapper[4733]: I1204 18:58:18.267289 4733 generic.go:334] "Generic (PLEG): container finished" podID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerID="4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021" exitCode=0 Dec 04 18:58:18 crc kubenswrapper[4733]: I1204 18:58:18.267386 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pkth4" event={"ID":"b8777330-05db-4780-9e3d-2f76a0b7a590","Type":"ContainerDied","Data":"4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021"} Dec 04 18:58:19 crc kubenswrapper[4733]: I1204 18:58:19.277266 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pkth4" event={"ID":"b8777330-05db-4780-9e3d-2f76a0b7a590","Type":"ContainerStarted","Data":"62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7"} Dec 04 18:58:19 crc kubenswrapper[4733]: I1204 18:58:19.297856 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pkth4" podStartSLOduration=1.827090476 podStartE2EDuration="4.297841755s" podCreationTimestamp="2025-12-04 18:58:15 +0000 UTC" firstStartedPulling="2025-12-04 18:58:16.22520168 +0000 UTC m=+4758.180562726" lastFinishedPulling="2025-12-04 18:58:18.695952959 +0000 UTC m=+4760.651314005" observedRunningTime="2025-12-04 18:58:19.295594485 +0000 UTC m=+4761.250955531" watchObservedRunningTime="2025-12-04 18:58:19.297841755 +0000 UTC m=+4761.253202801" Dec 04 18:58:20 crc kubenswrapper[4733]: I1204 18:58:20.673146 4733 scope.go:117] "RemoveContainer" containerID="0ed86f8827380f4b3fae7dd124841039d9a30de9d56cf47428336819a2069c00" Dec 04 18:58:25 crc kubenswrapper[4733]: I1204 18:58:25.546449 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:25 crc kubenswrapper[4733]: I1204 18:58:25.546996 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:25 crc kubenswrapper[4733]: I1204 18:58:25.597123 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:26 crc kubenswrapper[4733]: I1204 18:58:26.389877 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:26 crc kubenswrapper[4733]: I1204 18:58:26.439002 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pkth4"] Dec 04 18:58:28 crc kubenswrapper[4733]: I1204 18:58:28.361507 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pkth4" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerName="registry-server" containerID="cri-o://62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7" gracePeriod=2 Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.083227 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.087713 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-catalog-content\") pod \"b8777330-05db-4780-9e3d-2f76a0b7a590\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.088238 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-utilities\") pod \"b8777330-05db-4780-9e3d-2f76a0b7a590\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.089008 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-utilities" (OuterVolumeSpecName: "utilities") pod "b8777330-05db-4780-9e3d-2f76a0b7a590" (UID: "b8777330-05db-4780-9e3d-2f76a0b7a590"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.088271 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgtk9\" (UniqueName: \"kubernetes.io/projected/b8777330-05db-4780-9e3d-2f76a0b7a590-kube-api-access-lgtk9\") pod \"b8777330-05db-4780-9e3d-2f76a0b7a590\" (UID: \"b8777330-05db-4780-9e3d-2f76a0b7a590\") " Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.089243 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.095726 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8777330-05db-4780-9e3d-2f76a0b7a590-kube-api-access-lgtk9" (OuterVolumeSpecName: "kube-api-access-lgtk9") pod "b8777330-05db-4780-9e3d-2f76a0b7a590" (UID: "b8777330-05db-4780-9e3d-2f76a0b7a590"). InnerVolumeSpecName "kube-api-access-lgtk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.189935 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgtk9\" (UniqueName: \"kubernetes.io/projected/b8777330-05db-4780-9e3d-2f76a0b7a590-kube-api-access-lgtk9\") on node \"crc\" DevicePath \"\"" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.217870 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8777330-05db-4780-9e3d-2f76a0b7a590" (UID: "b8777330-05db-4780-9e3d-2f76a0b7a590"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.291112 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8777330-05db-4780-9e3d-2f76a0b7a590-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.383414 4733 generic.go:334] "Generic (PLEG): container finished" podID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerID="62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7" exitCode=0 Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.383455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pkth4" event={"ID":"b8777330-05db-4780-9e3d-2f76a0b7a590","Type":"ContainerDied","Data":"62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7"} Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.383483 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pkth4" event={"ID":"b8777330-05db-4780-9e3d-2f76a0b7a590","Type":"ContainerDied","Data":"69212896ba936a383614b842eb67f25d2db4ead12e49ddf5f4a0d8e7b52709d6"} Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.383499 4733 scope.go:117] "RemoveContainer" containerID="62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.383919 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pkth4" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.425623 4733 scope.go:117] "RemoveContainer" containerID="4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.425849 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pkth4"] Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.432162 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pkth4"] Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.453109 4733 scope.go:117] "RemoveContainer" containerID="b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.491997 4733 scope.go:117] "RemoveContainer" containerID="62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7" Dec 04 18:58:30 crc kubenswrapper[4733]: E1204 18:58:30.492565 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7\": container with ID starting with 62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7 not found: ID does not exist" containerID="62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.492851 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7"} err="failed to get container status \"62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7\": rpc error: code = NotFound desc = could not find container \"62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7\": container with ID starting with 62ce54a8fc512a40ef5de2938a5624072951baf2a89efd8ff1c198a927f72fc7 not found: ID does not exist" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.493093 4733 scope.go:117] "RemoveContainer" containerID="4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021" Dec 04 18:58:30 crc kubenswrapper[4733]: E1204 18:58:30.494428 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021\": container with ID starting with 4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021 not found: ID does not exist" containerID="4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.494463 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021"} err="failed to get container status \"4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021\": rpc error: code = NotFound desc = could not find container \"4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021\": container with ID starting with 4e4c38affaca34c732936a1413b8a4a619f0035be70d177d777e321ff6c61021 not found: ID does not exist" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.494480 4733 scope.go:117] "RemoveContainer" containerID="b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52" Dec 04 18:58:30 crc kubenswrapper[4733]: E1204 18:58:30.495198 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52\": container with ID starting with b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52 not found: ID does not exist" containerID="b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52" Dec 04 18:58:30 crc kubenswrapper[4733]: I1204 18:58:30.495252 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52"} err="failed to get container status \"b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52\": rpc error: code = NotFound desc = could not find container \"b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52\": container with ID starting with b918b19aa92bccedc43a163910b203eec054985868c3a05dadc4f00503d7ba52 not found: ID does not exist" Dec 04 18:58:32 crc kubenswrapper[4733]: I1204 18:58:32.344747 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" path="/var/lib/kubelet/pods/b8777330-05db-4780-9e3d-2f76a0b7a590/volumes" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.441278 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9wz4h"] Dec 04 18:58:44 crc kubenswrapper[4733]: E1204 18:58:44.442271 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerName="registry-server" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.442285 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerName="registry-server" Dec 04 18:58:44 crc kubenswrapper[4733]: E1204 18:58:44.442318 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerName="extract-content" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.442323 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerName="extract-content" Dec 04 18:58:44 crc kubenswrapper[4733]: E1204 18:58:44.442336 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerName="extract-utilities" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.442343 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerName="extract-utilities" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.442467 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8777330-05db-4780-9e3d-2f76a0b7a590" containerName="registry-server" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.443468 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.457348 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9wz4h"] Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.537392 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppxpg\" (UniqueName: \"kubernetes.io/projected/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-kube-api-access-ppxpg\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.538046 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-catalog-content\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.538180 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-utilities\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.639576 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppxpg\" (UniqueName: \"kubernetes.io/projected/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-kube-api-access-ppxpg\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.639724 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-catalog-content\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.639751 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-utilities\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.640412 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-utilities\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.640563 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-catalog-content\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.660180 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppxpg\" (UniqueName: \"kubernetes.io/projected/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-kube-api-access-ppxpg\") pod \"certified-operators-9wz4h\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:44 crc kubenswrapper[4733]: I1204 18:58:44.789126 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:45 crc kubenswrapper[4733]: I1204 18:58:45.332706 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9wz4h"] Dec 04 18:58:45 crc kubenswrapper[4733]: I1204 18:58:45.530667 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wz4h" event={"ID":"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0","Type":"ContainerStarted","Data":"4835aa497d0b67911f287ff361e81616469277d24a8acb0d75161a38072e2cfc"} Dec 04 18:58:46 crc kubenswrapper[4733]: I1204 18:58:46.545291 4733 generic.go:334] "Generic (PLEG): container finished" podID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerID="98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577" exitCode=0 Dec 04 18:58:46 crc kubenswrapper[4733]: I1204 18:58:46.545357 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wz4h" event={"ID":"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0","Type":"ContainerDied","Data":"98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577"} Dec 04 18:58:48 crc kubenswrapper[4733]: I1204 18:58:48.561431 4733 generic.go:334] "Generic (PLEG): container finished" podID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerID="45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe" exitCode=0 Dec 04 18:58:48 crc kubenswrapper[4733]: I1204 18:58:48.561480 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wz4h" event={"ID":"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0","Type":"ContainerDied","Data":"45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe"} Dec 04 18:58:49 crc kubenswrapper[4733]: I1204 18:58:49.572033 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wz4h" event={"ID":"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0","Type":"ContainerStarted","Data":"ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498"} Dec 04 18:58:49 crc kubenswrapper[4733]: I1204 18:58:49.592942 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9wz4h" podStartSLOduration=3.108561858 podStartE2EDuration="5.592927286s" podCreationTimestamp="2025-12-04 18:58:44 +0000 UTC" firstStartedPulling="2025-12-04 18:58:46.548226005 +0000 UTC m=+4788.503587051" lastFinishedPulling="2025-12-04 18:58:49.032591433 +0000 UTC m=+4790.987952479" observedRunningTime="2025-12-04 18:58:49.589074551 +0000 UTC m=+4791.544435597" watchObservedRunningTime="2025-12-04 18:58:49.592927286 +0000 UTC m=+4791.548288332" Dec 04 18:58:54 crc kubenswrapper[4733]: I1204 18:58:54.789293 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:54 crc kubenswrapper[4733]: I1204 18:58:54.789741 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:54 crc kubenswrapper[4733]: I1204 18:58:54.845837 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:55 crc kubenswrapper[4733]: I1204 18:58:55.670694 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:55 crc kubenswrapper[4733]: I1204 18:58:55.726340 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9wz4h"] Dec 04 18:58:57 crc kubenswrapper[4733]: I1204 18:58:57.644982 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9wz4h" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerName="registry-server" containerID="cri-o://ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498" gracePeriod=2 Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.097189 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.148463 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppxpg\" (UniqueName: \"kubernetes.io/projected/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-kube-api-access-ppxpg\") pod \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.148529 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-utilities\") pod \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.148661 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-catalog-content\") pod \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\" (UID: \"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0\") " Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.149646 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-utilities" (OuterVolumeSpecName: "utilities") pod "8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" (UID: "8590dc42-71a4-4d1c-94cc-4274b5b5a4f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.236609 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" (UID: "8590dc42-71a4-4d1c-94cc-4274b5b5a4f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.251071 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.251107 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.542893 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-kube-api-access-ppxpg" (OuterVolumeSpecName: "kube-api-access-ppxpg") pod "8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" (UID: "8590dc42-71a4-4d1c-94cc-4274b5b5a4f0"). InnerVolumeSpecName "kube-api-access-ppxpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.567164 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppxpg\" (UniqueName: \"kubernetes.io/projected/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0-kube-api-access-ppxpg\") on node \"crc\" DevicePath \"\"" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.656849 4733 generic.go:334] "Generic (PLEG): container finished" podID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerID="ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498" exitCode=0 Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.656913 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wz4h" event={"ID":"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0","Type":"ContainerDied","Data":"ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498"} Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.656972 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wz4h" event={"ID":"8590dc42-71a4-4d1c-94cc-4274b5b5a4f0","Type":"ContainerDied","Data":"4835aa497d0b67911f287ff361e81616469277d24a8acb0d75161a38072e2cfc"} Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.657005 4733 scope.go:117] "RemoveContainer" containerID="ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.657148 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wz4h" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.683537 4733 scope.go:117] "RemoveContainer" containerID="45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.713590 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9wz4h"] Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.723606 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9wz4h"] Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.731672 4733 scope.go:117] "RemoveContainer" containerID="98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.759977 4733 scope.go:117] "RemoveContainer" containerID="ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498" Dec 04 18:58:58 crc kubenswrapper[4733]: E1204 18:58:58.760404 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498\": container with ID starting with ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498 not found: ID does not exist" containerID="ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.760451 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498"} err="failed to get container status \"ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498\": rpc error: code = NotFound desc = could not find container \"ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498\": container with ID starting with ed7ba116a80ac8a68cef9139e7e33fb25d468cb0bf2d3f048dd17be57a8be498 not found: ID does not exist" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.760479 4733 scope.go:117] "RemoveContainer" containerID="45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe" Dec 04 18:58:58 crc kubenswrapper[4733]: E1204 18:58:58.761167 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe\": container with ID starting with 45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe not found: ID does not exist" containerID="45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.761222 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe"} err="failed to get container status \"45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe\": rpc error: code = NotFound desc = could not find container \"45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe\": container with ID starting with 45fd58df4d70c705288aa059bdade19d7f3bb8a225b3481bededf36877155cfe not found: ID does not exist" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.761284 4733 scope.go:117] "RemoveContainer" containerID="98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577" Dec 04 18:58:58 crc kubenswrapper[4733]: E1204 18:58:58.761616 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577\": container with ID starting with 98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577 not found: ID does not exist" containerID="98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577" Dec 04 18:58:58 crc kubenswrapper[4733]: I1204 18:58:58.761677 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577"} err="failed to get container status \"98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577\": rpc error: code = NotFound desc = could not find container \"98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577\": container with ID starting with 98d96bd8dcbd4517f68a854c93c49b48a89e127d965efdf21208d190356fe577 not found: ID does not exist" Dec 04 18:59:00 crc kubenswrapper[4733]: I1204 18:59:00.352485 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" path="/var/lib/kubelet/pods/8590dc42-71a4-4d1c-94cc-4274b5b5a4f0/volumes" Dec 04 18:59:45 crc kubenswrapper[4733]: I1204 18:59:45.362552 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 18:59:45 crc kubenswrapper[4733]: I1204 18:59:45.363363 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.142224 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7"] Dec 04 19:00:00 crc kubenswrapper[4733]: E1204 19:00:00.143027 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerName="extract-content" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.143040 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerName="extract-content" Dec 04 19:00:00 crc kubenswrapper[4733]: E1204 19:00:00.143069 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerName="registry-server" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.143075 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerName="registry-server" Dec 04 19:00:00 crc kubenswrapper[4733]: E1204 19:00:00.143089 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerName="extract-utilities" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.143095 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerName="extract-utilities" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.149092 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8590dc42-71a4-4d1c-94cc-4274b5b5a4f0" containerName="registry-server" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.149812 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.152947 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.153123 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.153203 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7"] Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.192129 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45ffb\" (UniqueName: \"kubernetes.io/projected/2beee50f-af17-48d4-9224-b34248a31cfd-kube-api-access-45ffb\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.192168 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beee50f-af17-48d4-9224-b34248a31cfd-config-volume\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.192190 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beee50f-af17-48d4-9224-b34248a31cfd-secret-volume\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.294020 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45ffb\" (UniqueName: \"kubernetes.io/projected/2beee50f-af17-48d4-9224-b34248a31cfd-kube-api-access-45ffb\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.294077 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beee50f-af17-48d4-9224-b34248a31cfd-config-volume\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.294115 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beee50f-af17-48d4-9224-b34248a31cfd-secret-volume\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.295032 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beee50f-af17-48d4-9224-b34248a31cfd-config-volume\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.307184 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beee50f-af17-48d4-9224-b34248a31cfd-secret-volume\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.316400 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45ffb\" (UniqueName: \"kubernetes.io/projected/2beee50f-af17-48d4-9224-b34248a31cfd-kube-api-access-45ffb\") pod \"collect-profiles-29414580-br8z7\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.472020 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:00 crc kubenswrapper[4733]: I1204 19:00:00.941706 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7"] Dec 04 19:00:01 crc kubenswrapper[4733]: I1204 19:00:01.235102 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" event={"ID":"2beee50f-af17-48d4-9224-b34248a31cfd","Type":"ContainerStarted","Data":"98399f88ed3c435260e57a8e3a6c633977de3439f9edb5927487c36eab94b7c7"} Dec 04 19:00:01 crc kubenswrapper[4733]: I1204 19:00:01.236542 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" event={"ID":"2beee50f-af17-48d4-9224-b34248a31cfd","Type":"ContainerStarted","Data":"102b4ebe8aa634069ee11cd4aaa1a80f9414e57215235598ca80d9e6923a8f0c"} Dec 04 19:00:02 crc kubenswrapper[4733]: I1204 19:00:02.245354 4733 generic.go:334] "Generic (PLEG): container finished" podID="2beee50f-af17-48d4-9224-b34248a31cfd" containerID="98399f88ed3c435260e57a8e3a6c633977de3439f9edb5927487c36eab94b7c7" exitCode=0 Dec 04 19:00:02 crc kubenswrapper[4733]: I1204 19:00:02.245416 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" event={"ID":"2beee50f-af17-48d4-9224-b34248a31cfd","Type":"ContainerDied","Data":"98399f88ed3c435260e57a8e3a6c633977de3439f9edb5927487c36eab94b7c7"} Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.593431 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.745625 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45ffb\" (UniqueName: \"kubernetes.io/projected/2beee50f-af17-48d4-9224-b34248a31cfd-kube-api-access-45ffb\") pod \"2beee50f-af17-48d4-9224-b34248a31cfd\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.745675 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beee50f-af17-48d4-9224-b34248a31cfd-secret-volume\") pod \"2beee50f-af17-48d4-9224-b34248a31cfd\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.745701 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beee50f-af17-48d4-9224-b34248a31cfd-config-volume\") pod \"2beee50f-af17-48d4-9224-b34248a31cfd\" (UID: \"2beee50f-af17-48d4-9224-b34248a31cfd\") " Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.746403 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2beee50f-af17-48d4-9224-b34248a31cfd-config-volume" (OuterVolumeSpecName: "config-volume") pod "2beee50f-af17-48d4-9224-b34248a31cfd" (UID: "2beee50f-af17-48d4-9224-b34248a31cfd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.751607 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2beee50f-af17-48d4-9224-b34248a31cfd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2beee50f-af17-48d4-9224-b34248a31cfd" (UID: "2beee50f-af17-48d4-9224-b34248a31cfd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.752688 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2beee50f-af17-48d4-9224-b34248a31cfd-kube-api-access-45ffb" (OuterVolumeSpecName: "kube-api-access-45ffb") pod "2beee50f-af17-48d4-9224-b34248a31cfd" (UID: "2beee50f-af17-48d4-9224-b34248a31cfd"). InnerVolumeSpecName "kube-api-access-45ffb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.847276 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45ffb\" (UniqueName: \"kubernetes.io/projected/2beee50f-af17-48d4-9224-b34248a31cfd-kube-api-access-45ffb\") on node \"crc\" DevicePath \"\"" Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.847332 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2beee50f-af17-48d4-9224-b34248a31cfd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 19:00:03 crc kubenswrapper[4733]: I1204 19:00:03.847353 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2beee50f-af17-48d4-9224-b34248a31cfd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 19:00:04 crc kubenswrapper[4733]: I1204 19:00:04.265455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" event={"ID":"2beee50f-af17-48d4-9224-b34248a31cfd","Type":"ContainerDied","Data":"102b4ebe8aa634069ee11cd4aaa1a80f9414e57215235598ca80d9e6923a8f0c"} Dec 04 19:00:04 crc kubenswrapper[4733]: I1204 19:00:04.265520 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="102b4ebe8aa634069ee11cd4aaa1a80f9414e57215235598ca80d9e6923a8f0c" Dec 04 19:00:04 crc kubenswrapper[4733]: I1204 19:00:04.265588 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7" Dec 04 19:00:04 crc kubenswrapper[4733]: I1204 19:00:04.347633 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr"] Dec 04 19:00:04 crc kubenswrapper[4733]: I1204 19:00:04.349111 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414535-cxqsr"] Dec 04 19:00:06 crc kubenswrapper[4733]: I1204 19:00:06.351547 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44e4ddcf-12df-4fad-bb44-96feea2c342e" path="/var/lib/kubelet/pods/44e4ddcf-12df-4fad-bb44-96feea2c342e/volumes" Dec 04 19:00:15 crc kubenswrapper[4733]: I1204 19:00:15.362141 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:00:15 crc kubenswrapper[4733]: I1204 19:00:15.364434 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:00:20 crc kubenswrapper[4733]: I1204 19:00:20.767127 4733 scope.go:117] "RemoveContainer" containerID="e3013392776988c35cf517b0fbbdfb7ad63bab4e41f8b5cb44dca286bf157f5b" Dec 04 19:00:45 crc kubenswrapper[4733]: I1204 19:00:45.362348 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:00:45 crc kubenswrapper[4733]: I1204 19:00:45.363026 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:00:45 crc kubenswrapper[4733]: I1204 19:00:45.363075 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:00:45 crc kubenswrapper[4733]: I1204 19:00:45.364067 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"133576a4fa20da80bc0ad5a9ec576a91784d7d1b185b32466a081bc4c9e98f7f"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:00:45 crc kubenswrapper[4733]: I1204 19:00:45.364180 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://133576a4fa20da80bc0ad5a9ec576a91784d7d1b185b32466a081bc4c9e98f7f" gracePeriod=600 Dec 04 19:00:45 crc kubenswrapper[4733]: I1204 19:00:45.669560 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="133576a4fa20da80bc0ad5a9ec576a91784d7d1b185b32466a081bc4c9e98f7f" exitCode=0 Dec 04 19:00:45 crc kubenswrapper[4733]: I1204 19:00:45.669626 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"133576a4fa20da80bc0ad5a9ec576a91784d7d1b185b32466a081bc4c9e98f7f"} Dec 04 19:00:45 crc kubenswrapper[4733]: I1204 19:00:45.670122 4733 scope.go:117] "RemoveContainer" containerID="39869b10e1cd0b03600c58de0453984d34298b19452143ae53b846166ab7813a" Dec 04 19:00:46 crc kubenswrapper[4733]: I1204 19:00:46.679860 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf"} Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.491423 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76d8c4d77f-2hj2q"] Dec 04 19:00:52 crc kubenswrapper[4733]: E1204 19:00:52.492123 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2beee50f-af17-48d4-9224-b34248a31cfd" containerName="collect-profiles" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.492135 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2beee50f-af17-48d4-9224-b34248a31cfd" containerName="collect-profiles" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.492302 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2beee50f-af17-48d4-9224-b34248a31cfd" containerName="collect-profiles" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.492958 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.494991 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.495174 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.495222 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.495506 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.495592 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-nm75n" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.503354 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfwnf\" (UniqueName: \"kubernetes.io/projected/7872710d-9e81-4f34-b408-1aa8c48d1007-kube-api-access-pfwnf\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.503434 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-dns-svc\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.503499 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-config\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.509136 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76d8c4d77f-2hj2q"] Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.604556 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-config\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.604630 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfwnf\" (UniqueName: \"kubernetes.io/projected/7872710d-9e81-4f34-b408-1aa8c48d1007-kube-api-access-pfwnf\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.604664 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-dns-svc\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.605589 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-dns-svc\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.605781 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-config\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.625004 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfwnf\" (UniqueName: \"kubernetes.io/projected/7872710d-9e81-4f34-b408-1aa8c48d1007-kube-api-access-pfwnf\") pod \"dnsmasq-dns-76d8c4d77f-2hj2q\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.785942 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cbb4f659c-sbt7t"] Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.787703 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.811161 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.847208 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cbb4f659c-sbt7t"] Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.909108 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fr8z\" (UniqueName: \"kubernetes.io/projected/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-kube-api-access-4fr8z\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.909297 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-config\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:52 crc kubenswrapper[4733]: I1204 19:00:52.909331 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-dns-svc\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.010822 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-config\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.010863 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-dns-svc\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.010922 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fr8z\" (UniqueName: \"kubernetes.io/projected/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-kube-api-access-4fr8z\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.012207 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-config\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.012704 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-dns-svc\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.050736 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fr8z\" (UniqueName: \"kubernetes.io/projected/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-kube-api-access-4fr8z\") pod \"dnsmasq-dns-7cbb4f659c-sbt7t\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.105136 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.109417 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76d8c4d77f-2hj2q"] Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.525411 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cbb4f659c-sbt7t"] Dec 04 19:00:53 crc kubenswrapper[4733]: W1204 19:00:53.527743 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ae9da60_b365_4932_87d6_6bc9fe29d2cb.slice/crio-52f3d1c89f0bc68079cca98bf6370d6b6dd9e68f9fb252bf1c1e4c03a2ef1849 WatchSource:0}: Error finding container 52f3d1c89f0bc68079cca98bf6370d6b6dd9e68f9fb252bf1c1e4c03a2ef1849: Status 404 returned error can't find the container with id 52f3d1c89f0bc68079cca98bf6370d6b6dd9e68f9fb252bf1c1e4c03a2ef1849 Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.678132 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.680009 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.681826 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.682203 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.682385 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.682637 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-4hz4z" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.682813 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.706937 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.740515 4733 generic.go:334] "Generic (PLEG): container finished" podID="7872710d-9e81-4f34-b408-1aa8c48d1007" containerID="603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c" exitCode=0 Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.740580 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" event={"ID":"7872710d-9e81-4f34-b408-1aa8c48d1007","Type":"ContainerDied","Data":"603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c"} Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.740606 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" event={"ID":"7872710d-9e81-4f34-b408-1aa8c48d1007","Type":"ContainerStarted","Data":"feb1bb443550299d0a88908859d5db4902a0d9289ae3aa91305e9ccd281e0c5e"} Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.741977 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" containerID="2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b" exitCode=0 Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.742011 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" event={"ID":"8ae9da60-b365-4932-87d6-6bc9fe29d2cb","Type":"ContainerDied","Data":"2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b"} Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.742034 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" event={"ID":"8ae9da60-b365-4932-87d6-6bc9fe29d2cb","Type":"ContainerStarted","Data":"52f3d1c89f0bc68079cca98bf6370d6b6dd9e68f9fb252bf1c1e4c03a2ef1849"} Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.822897 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.822947 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.822967 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.822984 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.822998 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc170837-84b1-4112-b0fc-e483e9df9f60-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.823018 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc170837-84b1-4112-b0fc-e483e9df9f60-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.823045 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.823065 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghm78\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-kube-api-access-ghm78\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.823088 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924230 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924293 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924329 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924354 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924373 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc170837-84b1-4112-b0fc-e483e9df9f60-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924411 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc170837-84b1-4112-b0fc-e483e9df9f60-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924445 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924473 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghm78\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-kube-api-access-ghm78\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.924506 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.925485 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.925628 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.925895 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.926418 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.928352 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.928694 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.928734 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bff92f9126deea5cf8426ea124f9832b721e3fe298f7a31ffdaeaf0c710d38b4/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.929005 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc170837-84b1-4112-b0fc-e483e9df9f60-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.934279 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc170837-84b1-4112-b0fc-e483e9df9f60-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.947854 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.949393 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.950664 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghm78\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-kube-api-access-ghm78\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.953708 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.953963 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.954066 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.953971 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.954322 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gpm72" Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.965253 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:00:53 crc kubenswrapper[4733]: I1204 19:00:53.997500 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"rabbitmq-server-0\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " pod="openstack/rabbitmq-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: E1204 19:00:54.053535 4733 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 04 19:00:54 crc kubenswrapper[4733]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/7872710d-9e81-4f34-b408-1aa8c48d1007/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 04 19:00:54 crc kubenswrapper[4733]: > podSandboxID="feb1bb443550299d0a88908859d5db4902a0d9289ae3aa91305e9ccd281e0c5e" Dec 04 19:00:54 crc kubenswrapper[4733]: E1204 19:00:54.053720 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 19:00:54 crc kubenswrapper[4733]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pfwnf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-76d8c4d77f-2hj2q_openstack(7872710d-9e81-4f34-b408-1aa8c48d1007): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/7872710d-9e81-4f34-b408-1aa8c48d1007/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 04 19:00:54 crc kubenswrapper[4733]: > logger="UnhandledError" Dec 04 19:00:54 crc kubenswrapper[4733]: E1204 19:00:54.056332 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/7872710d-9e81-4f34-b408-1aa8c48d1007/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" podUID="7872710d-9e81-4f34-b408-1aa8c48d1007" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.127245 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvmsd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-kube-api-access-tvmsd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.127538 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.127651 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.127761 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.127918 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.128033 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fea0eafe-f021-49e1-b949-450a65b297db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.128150 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fea0eafe-f021-49e1-b949-450a65b297db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.128270 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.128391 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.145040 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.229906 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.229956 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.230002 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.230030 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fea0eafe-f021-49e1-b949-450a65b297db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.230386 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.230056 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fea0eafe-f021-49e1-b949-450a65b297db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.230854 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.230907 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.230944 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvmsd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-kube-api-access-tvmsd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.230981 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.231307 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.231633 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.232199 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.234646 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fea0eafe-f021-49e1-b949-450a65b297db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.235040 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.235921 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.235975 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2a402ac0835941d16d1fb7a83ff5ff9713c626041a7644231b79bfaf8135d6ab/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.237880 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fea0eafe-f021-49e1-b949-450a65b297db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.251384 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvmsd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-kube-api-access-tvmsd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.264412 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"rabbitmq-cell1-server-0\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.417805 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.558553 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:00:54 crc kubenswrapper[4733]: W1204 19:00:54.576178 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc170837_84b1_4112_b0fc_e483e9df9f60.slice/crio-ffb3459ab3e46a3142910e71e264cb94dfbae4edd7c2055488283afc2ef38712 WatchSource:0}: Error finding container ffb3459ab3e46a3142910e71e264cb94dfbae4edd7c2055488283afc2ef38712: Status 404 returned error can't find the container with id ffb3459ab3e46a3142910e71e264cb94dfbae4edd7c2055488283afc2ef38712 Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.753491 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc170837-84b1-4112-b0fc-e483e9df9f60","Type":"ContainerStarted","Data":"ffb3459ab3e46a3142910e71e264cb94dfbae4edd7c2055488283afc2ef38712"} Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.756062 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" event={"ID":"8ae9da60-b365-4932-87d6-6bc9fe29d2cb","Type":"ContainerStarted","Data":"56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233"} Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.756394 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.796640 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" podStartSLOduration=2.796619858 podStartE2EDuration="2.796619858s" podCreationTimestamp="2025-12-04 19:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:00:54.795411306 +0000 UTC m=+4916.750772362" watchObservedRunningTime="2025-12-04 19:00:54.796619858 +0000 UTC m=+4916.751980904" Dec 04 19:00:54 crc kubenswrapper[4733]: I1204 19:00:54.853196 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:00:54 crc kubenswrapper[4733]: W1204 19:00:54.857272 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfea0eafe_f021_49e1_b949_450a65b297db.slice/crio-625a075da08bebed457b7094108e5851b6fd71f8379af91c0d24052bec4ac93d WatchSource:0}: Error finding container 625a075da08bebed457b7094108e5851b6fd71f8379af91c0d24052bec4ac93d: Status 404 returned error can't find the container with id 625a075da08bebed457b7094108e5851b6fd71f8379af91c0d24052bec4ac93d Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.158388 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.160611 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.163152 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.164612 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.169126 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-497q6" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.170271 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.176871 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.179521 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.246503 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-kolla-config\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.246618 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-config-data-default\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.246657 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk6lw\" (UniqueName: \"kubernetes.io/projected/e5c74c10-5554-486b-b79d-af7a2fa49019-kube-api-access-tk6lw\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.246725 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.246785 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-36f4d1d7-c084-48c2-b2f9-6978b8a2d48c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-36f4d1d7-c084-48c2-b2f9-6978b8a2d48c\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.246862 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5c74c10-5554-486b-b79d-af7a2fa49019-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.246898 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c74c10-5554-486b-b79d-af7a2fa49019-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.246940 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5c74c10-5554-486b-b79d-af7a2fa49019-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.348256 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.348334 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-36f4d1d7-c084-48c2-b2f9-6978b8a2d48c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-36f4d1d7-c084-48c2-b2f9-6978b8a2d48c\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.348379 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5c74c10-5554-486b-b79d-af7a2fa49019-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.348406 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c74c10-5554-486b-b79d-af7a2fa49019-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.348440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5c74c10-5554-486b-b79d-af7a2fa49019-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.348475 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-kolla-config\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.348541 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-config-data-default\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.348575 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk6lw\" (UniqueName: \"kubernetes.io/projected/e5c74c10-5554-486b-b79d-af7a2fa49019-kube-api-access-tk6lw\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.350625 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.351971 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-kolla-config\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.353741 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e5c74c10-5554-486b-b79d-af7a2fa49019-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.353830 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e5c74c10-5554-486b-b79d-af7a2fa49019-config-data-default\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.358198 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5c74c10-5554-486b-b79d-af7a2fa49019-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.358231 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c74c10-5554-486b-b79d-af7a2fa49019-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.359132 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.359160 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-36f4d1d7-c084-48c2-b2f9-6978b8a2d48c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-36f4d1d7-c084-48c2-b2f9-6978b8a2d48c\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8d59cdf59d0b0282d96cc72a08c7703cd29e4067210c0086557f1f225c6ce17f/globalmount\"" pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.370188 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk6lw\" (UniqueName: \"kubernetes.io/projected/e5c74c10-5554-486b-b79d-af7a2fa49019-kube-api-access-tk6lw\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.393283 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-36f4d1d7-c084-48c2-b2f9-6978b8a2d48c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-36f4d1d7-c084-48c2-b2f9-6978b8a2d48c\") pod \"openstack-galera-0\" (UID: \"e5c74c10-5554-486b-b79d-af7a2fa49019\") " pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.492423 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.547191 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.548057 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.550183 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.550311 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-crznl" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.571421 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.665565 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9crnf\" (UniqueName: \"kubernetes.io/projected/b365fa5f-abce-4758-a4f2-1307febe0243-kube-api-access-9crnf\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.666104 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b365fa5f-abce-4758-a4f2-1307febe0243-config-data\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.666452 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b365fa5f-abce-4758-a4f2-1307febe0243-kolla-config\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.777717 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b365fa5f-abce-4758-a4f2-1307febe0243-kolla-config\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.777905 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9crnf\" (UniqueName: \"kubernetes.io/projected/b365fa5f-abce-4758-a4f2-1307febe0243-kube-api-access-9crnf\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.777959 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b365fa5f-abce-4758-a4f2-1307febe0243-config-data\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.778958 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b365fa5f-abce-4758-a4f2-1307febe0243-config-data\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.779495 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b365fa5f-abce-4758-a4f2-1307febe0243-kolla-config\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.781515 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" event={"ID":"7872710d-9e81-4f34-b408-1aa8c48d1007","Type":"ContainerStarted","Data":"79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4"} Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.782517 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.789603 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fea0eafe-f021-49e1-b949-450a65b297db","Type":"ContainerStarted","Data":"625a075da08bebed457b7094108e5851b6fd71f8379af91c0d24052bec4ac93d"} Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.816138 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" podStartSLOduration=3.81612075 podStartE2EDuration="3.81612075s" podCreationTimestamp="2025-12-04 19:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:00:55.811289319 +0000 UTC m=+4917.766650365" watchObservedRunningTime="2025-12-04 19:00:55.81612075 +0000 UTC m=+4917.771481796" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.845246 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9crnf\" (UniqueName: \"kubernetes.io/projected/b365fa5f-abce-4758-a4f2-1307febe0243-kube-api-access-9crnf\") pod \"memcached-0\" (UID: \"b365fa5f-abce-4758-a4f2-1307febe0243\") " pod="openstack/memcached-0" Dec 04 19:00:55 crc kubenswrapper[4733]: I1204 19:00:55.872184 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.045029 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.403385 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 19:00:56 crc kubenswrapper[4733]: W1204 19:00:56.405361 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb365fa5f_abce_4758_a4f2_1307febe0243.slice/crio-a90a04d0e91850f05dd9f0a579d08765fa492c306fc122b8cff1896f2214e59f WatchSource:0}: Error finding container a90a04d0e91850f05dd9f0a579d08765fa492c306fc122b8cff1896f2214e59f: Status 404 returned error can't find the container with id a90a04d0e91850f05dd9f0a579d08765fa492c306fc122b8cff1896f2214e59f Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.752903 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.754466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.757636 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.757711 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.757636 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.758885 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-69cjs" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.771690 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.809827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b365fa5f-abce-4758-a4f2-1307febe0243","Type":"ContainerStarted","Data":"fafdf3dd081367f7d6d2f42c8a9c5f733ae5ed3f5a7421c29912d341a955a7a6"} Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.809885 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b365fa5f-abce-4758-a4f2-1307febe0243","Type":"ContainerStarted","Data":"a90a04d0e91850f05dd9f0a579d08765fa492c306fc122b8cff1896f2214e59f"} Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.809933 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.814081 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fea0eafe-f021-49e1-b949-450a65b297db","Type":"ContainerStarted","Data":"b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816"} Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.816383 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc170837-84b1-4112-b0fc-e483e9df9f60","Type":"ContainerStarted","Data":"db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e"} Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.819753 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e5c74c10-5554-486b-b79d-af7a2fa49019","Type":"ContainerStarted","Data":"bb161c26b6e80b585f9d74842f7f91bc496626a052630a435114e6570e3ebdd7"} Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.819827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e5c74c10-5554-486b-b79d-af7a2fa49019","Type":"ContainerStarted","Data":"21f8ab510b6a413702818156c1b8f018aae1131b5a5288b795835dd60dfa7531"} Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.838143 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.8381251189999999 podStartE2EDuration="1.838125119s" podCreationTimestamp="2025-12-04 19:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:00:56.835400455 +0000 UTC m=+4918.790761501" watchObservedRunningTime="2025-12-04 19:00:56.838125119 +0000 UTC m=+4918.793486165" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.904203 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f7133c01-47a1-4adc-9d8e-2a314124aa03-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.904290 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrvzj\" (UniqueName: \"kubernetes.io/projected/f7133c01-47a1-4adc-9d8e-2a314124aa03-kube-api-access-lrvzj\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.904334 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ce74ce38-38bf-4665-be0e-7471ffc05d96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ce74ce38-38bf-4665-be0e-7471ffc05d96\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.904428 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.904522 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.905410 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7133c01-47a1-4adc-9d8e-2a314124aa03-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.905486 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:56 crc kubenswrapper[4733]: I1204 19:00:56.905517 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7133c01-47a1-4adc-9d8e-2a314124aa03-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.007200 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7133c01-47a1-4adc-9d8e-2a314124aa03-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.007285 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.007319 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7133c01-47a1-4adc-9d8e-2a314124aa03-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.007440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f7133c01-47a1-4adc-9d8e-2a314124aa03-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.007459 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrvzj\" (UniqueName: \"kubernetes.io/projected/f7133c01-47a1-4adc-9d8e-2a314124aa03-kube-api-access-lrvzj\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.007478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ce74ce38-38bf-4665-be0e-7471ffc05d96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ce74ce38-38bf-4665-be0e-7471ffc05d96\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.007513 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.007533 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.008235 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.009211 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f7133c01-47a1-4adc-9d8e-2a314124aa03-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.009711 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.010440 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7133c01-47a1-4adc-9d8e-2a314124aa03-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.014790 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.014872 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ce74ce38-38bf-4665-be0e-7471ffc05d96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ce74ce38-38bf-4665-be0e-7471ffc05d96\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8945e2db61ba17131830c3feb22ab54ece06adf1fb2d967386ddc68ee43a1d16/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.042086 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7133c01-47a1-4adc-9d8e-2a314124aa03-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.043251 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7133c01-47a1-4adc-9d8e-2a314124aa03-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.052096 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrvzj\" (UniqueName: \"kubernetes.io/projected/f7133c01-47a1-4adc-9d8e-2a314124aa03-kube-api-access-lrvzj\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.059313 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ce74ce38-38bf-4665-be0e-7471ffc05d96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ce74ce38-38bf-4665-be0e-7471ffc05d96\") pod \"openstack-cell1-galera-0\" (UID: \"f7133c01-47a1-4adc-9d8e-2a314124aa03\") " pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.115830 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 19:00:57 crc kubenswrapper[4733]: W1204 19:00:57.615614 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7133c01_47a1_4adc_9d8e_2a314124aa03.slice/crio-54243cdcceb5d8b8d1c0812b374a452786cf9b85701193946814aa9be0e7fdb0 WatchSource:0}: Error finding container 54243cdcceb5d8b8d1c0812b374a452786cf9b85701193946814aa9be0e7fdb0: Status 404 returned error can't find the container with id 54243cdcceb5d8b8d1c0812b374a452786cf9b85701193946814aa9be0e7fdb0 Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.624515 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 19:00:57 crc kubenswrapper[4733]: I1204 19:00:57.835410 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f7133c01-47a1-4adc-9d8e-2a314124aa03","Type":"ContainerStarted","Data":"54243cdcceb5d8b8d1c0812b374a452786cf9b85701193946814aa9be0e7fdb0"} Dec 04 19:00:58 crc kubenswrapper[4733]: I1204 19:00:58.843640 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f7133c01-47a1-4adc-9d8e-2a314124aa03","Type":"ContainerStarted","Data":"490eb935818aeded4d6d22e37b72e03f222556fb6d625831212bccd736126b7b"} Dec 04 19:00:59 crc kubenswrapper[4733]: I1204 19:00:59.857983 4733 generic.go:334] "Generic (PLEG): container finished" podID="e5c74c10-5554-486b-b79d-af7a2fa49019" containerID="bb161c26b6e80b585f9d74842f7f91bc496626a052630a435114e6570e3ebdd7" exitCode=0 Dec 04 19:00:59 crc kubenswrapper[4733]: I1204 19:00:59.858101 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e5c74c10-5554-486b-b79d-af7a2fa49019","Type":"ContainerDied","Data":"bb161c26b6e80b585f9d74842f7f91bc496626a052630a435114e6570e3ebdd7"} Dec 04 19:01:00 crc kubenswrapper[4733]: I1204 19:01:00.869714 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e5c74c10-5554-486b-b79d-af7a2fa49019","Type":"ContainerStarted","Data":"5c5bdbfd8202015fc8a14a64567801f57b3a70be6d902a083baf36c59a4e5dd9"} Dec 04 19:01:00 crc kubenswrapper[4733]: I1204 19:01:00.918767 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=6.91861421 podStartE2EDuration="6.91861421s" podCreationTimestamp="2025-12-04 19:00:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:01:00.903169952 +0000 UTC m=+4922.858531038" watchObservedRunningTime="2025-12-04 19:01:00.91861421 +0000 UTC m=+4922.873975346" Dec 04 19:01:01 crc kubenswrapper[4733]: I1204 19:01:01.881867 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f7133c01-47a1-4adc-9d8e-2a314124aa03","Type":"ContainerDied","Data":"490eb935818aeded4d6d22e37b72e03f222556fb6d625831212bccd736126b7b"} Dec 04 19:01:01 crc kubenswrapper[4733]: I1204 19:01:01.882136 4733 generic.go:334] "Generic (PLEG): container finished" podID="f7133c01-47a1-4adc-9d8e-2a314124aa03" containerID="490eb935818aeded4d6d22e37b72e03f222556fb6d625831212bccd736126b7b" exitCode=0 Dec 04 19:01:02 crc kubenswrapper[4733]: I1204 19:01:02.813035 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:01:02 crc kubenswrapper[4733]: I1204 19:01:02.912956 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f7133c01-47a1-4adc-9d8e-2a314124aa03","Type":"ContainerStarted","Data":"686168c6012363d992dfc46088d0138a3a34d96873e711a50b89163f4e73efad"} Dec 04 19:01:02 crc kubenswrapper[4733]: I1204 19:01:02.940258 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.940238682 podStartE2EDuration="7.940238682s" podCreationTimestamp="2025-12-04 19:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:01:02.939210385 +0000 UTC m=+4924.894571481" watchObservedRunningTime="2025-12-04 19:01:02.940238682 +0000 UTC m=+4924.895599728" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.107285 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.159253 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76d8c4d77f-2hj2q"] Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.159469 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" podUID="7872710d-9e81-4f34-b408-1aa8c48d1007" containerName="dnsmasq-dns" containerID="cri-o://79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4" gracePeriod=10 Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.561002 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.648955 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-config\") pod \"7872710d-9e81-4f34-b408-1aa8c48d1007\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.649034 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-dns-svc\") pod \"7872710d-9e81-4f34-b408-1aa8c48d1007\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.649183 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfwnf\" (UniqueName: \"kubernetes.io/projected/7872710d-9e81-4f34-b408-1aa8c48d1007-kube-api-access-pfwnf\") pod \"7872710d-9e81-4f34-b408-1aa8c48d1007\" (UID: \"7872710d-9e81-4f34-b408-1aa8c48d1007\") " Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.656623 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7872710d-9e81-4f34-b408-1aa8c48d1007-kube-api-access-pfwnf" (OuterVolumeSpecName: "kube-api-access-pfwnf") pod "7872710d-9e81-4f34-b408-1aa8c48d1007" (UID: "7872710d-9e81-4f34-b408-1aa8c48d1007"). InnerVolumeSpecName "kube-api-access-pfwnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.691570 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7872710d-9e81-4f34-b408-1aa8c48d1007" (UID: "7872710d-9e81-4f34-b408-1aa8c48d1007"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.696973 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-config" (OuterVolumeSpecName: "config") pod "7872710d-9e81-4f34-b408-1aa8c48d1007" (UID: "7872710d-9e81-4f34-b408-1aa8c48d1007"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.750916 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.750947 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7872710d-9e81-4f34-b408-1aa8c48d1007-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.750956 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfwnf\" (UniqueName: \"kubernetes.io/projected/7872710d-9e81-4f34-b408-1aa8c48d1007-kube-api-access-pfwnf\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.922475 4733 generic.go:334] "Generic (PLEG): container finished" podID="7872710d-9e81-4f34-b408-1aa8c48d1007" containerID="79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4" exitCode=0 Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.922548 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.922536 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" event={"ID":"7872710d-9e81-4f34-b408-1aa8c48d1007","Type":"ContainerDied","Data":"79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4"} Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.922854 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d8c4d77f-2hj2q" event={"ID":"7872710d-9e81-4f34-b408-1aa8c48d1007","Type":"ContainerDied","Data":"feb1bb443550299d0a88908859d5db4902a0d9289ae3aa91305e9ccd281e0c5e"} Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.922874 4733 scope.go:117] "RemoveContainer" containerID="79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.957247 4733 scope.go:117] "RemoveContainer" containerID="603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.968757 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76d8c4d77f-2hj2q"] Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.974659 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76d8c4d77f-2hj2q"] Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.979081 4733 scope.go:117] "RemoveContainer" containerID="79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4" Dec 04 19:01:03 crc kubenswrapper[4733]: E1204 19:01:03.979619 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4\": container with ID starting with 79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4 not found: ID does not exist" containerID="79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.979662 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4"} err="failed to get container status \"79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4\": rpc error: code = NotFound desc = could not find container \"79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4\": container with ID starting with 79b4088d0213b42da9dbb59086268967ccd05b8c999fdfe962f965598062c5b4 not found: ID does not exist" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.979688 4733 scope.go:117] "RemoveContainer" containerID="603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c" Dec 04 19:01:03 crc kubenswrapper[4733]: E1204 19:01:03.980055 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c\": container with ID starting with 603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c not found: ID does not exist" containerID="603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c" Dec 04 19:01:03 crc kubenswrapper[4733]: I1204 19:01:03.980079 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c"} err="failed to get container status \"603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c\": rpc error: code = NotFound desc = could not find container \"603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c\": container with ID starting with 603c7ccd9e1c55acf285d606832ac4282adf2dc93ca58c9a5cebd377697a900c not found: ID does not exist" Dec 04 19:01:04 crc kubenswrapper[4733]: I1204 19:01:04.353004 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7872710d-9e81-4f34-b408-1aa8c48d1007" path="/var/lib/kubelet/pods/7872710d-9e81-4f34-b408-1aa8c48d1007/volumes" Dec 04 19:01:05 crc kubenswrapper[4733]: I1204 19:01:05.493747 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 04 19:01:05 crc kubenswrapper[4733]: I1204 19:01:05.493829 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 04 19:01:05 crc kubenswrapper[4733]: I1204 19:01:05.874012 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 04 19:01:07 crc kubenswrapper[4733]: I1204 19:01:07.117561 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 04 19:01:07 crc kubenswrapper[4733]: I1204 19:01:07.118647 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 04 19:01:07 crc kubenswrapper[4733]: I1204 19:01:07.727861 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 04 19:01:07 crc kubenswrapper[4733]: I1204 19:01:07.805441 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 04 19:01:09 crc kubenswrapper[4733]: I1204 19:01:09.338118 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 04 19:01:09 crc kubenswrapper[4733]: I1204 19:01:09.449264 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 04 19:01:28 crc kubenswrapper[4733]: I1204 19:01:28.143983 4733 generic.go:334] "Generic (PLEG): container finished" podID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerID="db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e" exitCode=0 Dec 04 19:01:28 crc kubenswrapper[4733]: I1204 19:01:28.144092 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc170837-84b1-4112-b0fc-e483e9df9f60","Type":"ContainerDied","Data":"db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e"} Dec 04 19:01:29 crc kubenswrapper[4733]: I1204 19:01:29.153251 4733 generic.go:334] "Generic (PLEG): container finished" podID="fea0eafe-f021-49e1-b949-450a65b297db" containerID="b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816" exitCode=0 Dec 04 19:01:29 crc kubenswrapper[4733]: I1204 19:01:29.153360 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fea0eafe-f021-49e1-b949-450a65b297db","Type":"ContainerDied","Data":"b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816"} Dec 04 19:01:29 crc kubenswrapper[4733]: I1204 19:01:29.155649 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc170837-84b1-4112-b0fc-e483e9df9f60","Type":"ContainerStarted","Data":"4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a"} Dec 04 19:01:29 crc kubenswrapper[4733]: I1204 19:01:29.155891 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 19:01:29 crc kubenswrapper[4733]: I1204 19:01:29.207880 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.207865571 podStartE2EDuration="37.207865571s" podCreationTimestamp="2025-12-04 19:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:01:29.205971819 +0000 UTC m=+4951.161332875" watchObservedRunningTime="2025-12-04 19:01:29.207865571 +0000 UTC m=+4951.163226607" Dec 04 19:01:30 crc kubenswrapper[4733]: I1204 19:01:30.167810 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fea0eafe-f021-49e1-b949-450a65b297db","Type":"ContainerStarted","Data":"c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e"} Dec 04 19:01:30 crc kubenswrapper[4733]: I1204 19:01:30.168520 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:01:30 crc kubenswrapper[4733]: I1204 19:01:30.199428 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.199384285 podStartE2EDuration="38.199384285s" podCreationTimestamp="2025-12-04 19:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:01:30.187318238 +0000 UTC m=+4952.142679304" watchObservedRunningTime="2025-12-04 19:01:30.199384285 +0000 UTC m=+4952.154745341" Dec 04 19:01:44 crc kubenswrapper[4733]: I1204 19:01:44.148710 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 19:01:44 crc kubenswrapper[4733]: I1204 19:01:44.420814 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:01:49 crc kubenswrapper[4733]: I1204 19:01:49.918363 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f79bf7859-x6fqr"] Dec 04 19:01:49 crc kubenswrapper[4733]: E1204 19:01:49.919230 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7872710d-9e81-4f34-b408-1aa8c48d1007" containerName="dnsmasq-dns" Dec 04 19:01:49 crc kubenswrapper[4733]: I1204 19:01:49.919245 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7872710d-9e81-4f34-b408-1aa8c48d1007" containerName="dnsmasq-dns" Dec 04 19:01:49 crc kubenswrapper[4733]: E1204 19:01:49.919282 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7872710d-9e81-4f34-b408-1aa8c48d1007" containerName="init" Dec 04 19:01:49 crc kubenswrapper[4733]: I1204 19:01:49.919290 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7872710d-9e81-4f34-b408-1aa8c48d1007" containerName="init" Dec 04 19:01:49 crc kubenswrapper[4733]: I1204 19:01:49.919470 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7872710d-9e81-4f34-b408-1aa8c48d1007" containerName="dnsmasq-dns" Dec 04 19:01:49 crc kubenswrapper[4733]: I1204 19:01:49.920439 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:49 crc kubenswrapper[4733]: I1204 19:01:49.959344 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f79bf7859-x6fqr"] Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.098790 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw2tf\" (UniqueName: \"kubernetes.io/projected/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-kube-api-access-zw2tf\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.099138 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-dns-svc\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.099178 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-config\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.201060 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-config\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.201233 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw2tf\" (UniqueName: \"kubernetes.io/projected/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-kube-api-access-zw2tf\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.201260 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-dns-svc\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.202333 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-dns-svc\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.202873 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-config\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.244598 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw2tf\" (UniqueName: \"kubernetes.io/projected/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-kube-api-access-zw2tf\") pod \"dnsmasq-dns-f79bf7859-x6fqr\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.254045 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.559451 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:01:50 crc kubenswrapper[4733]: I1204 19:01:50.803777 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f79bf7859-x6fqr"] Dec 04 19:01:51 crc kubenswrapper[4733]: I1204 19:01:51.372917 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" containerID="938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59" exitCode=0 Dec 04 19:01:51 crc kubenswrapper[4733]: I1204 19:01:51.373029 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" event={"ID":"2d3818e0-89ec-4cfd-8bd1-918c99424ae0","Type":"ContainerDied","Data":"938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59"} Dec 04 19:01:51 crc kubenswrapper[4733]: I1204 19:01:51.373279 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" event={"ID":"2d3818e0-89ec-4cfd-8bd1-918c99424ae0","Type":"ContainerStarted","Data":"2579a7d57539774294db6414d653b1ce2d17db92b523234a7ecbe697745a794a"} Dec 04 19:01:51 crc kubenswrapper[4733]: I1204 19:01:51.508313 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:01:52 crc kubenswrapper[4733]: I1204 19:01:52.382991 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" event={"ID":"2d3818e0-89ec-4cfd-8bd1-918c99424ae0","Type":"ContainerStarted","Data":"ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49"} Dec 04 19:01:52 crc kubenswrapper[4733]: I1204 19:01:52.383246 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:01:52 crc kubenswrapper[4733]: I1204 19:01:52.410359 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" podStartSLOduration=3.410344074 podStartE2EDuration="3.410344074s" podCreationTimestamp="2025-12-04 19:01:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:01:52.408766362 +0000 UTC m=+4974.364127418" watchObservedRunningTime="2025-12-04 19:01:52.410344074 +0000 UTC m=+4974.365705110" Dec 04 19:01:52 crc kubenswrapper[4733]: I1204 19:01:52.620851 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerName="rabbitmq" containerID="cri-o://4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a" gracePeriod=604798 Dec 04 19:01:53 crc kubenswrapper[4733]: I1204 19:01:53.309366 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="fea0eafe-f021-49e1-b949-450a65b297db" containerName="rabbitmq" containerID="cri-o://c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e" gracePeriod=604799 Dec 04 19:01:54 crc kubenswrapper[4733]: I1204 19:01:54.145916 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.244:5672: connect: connection refused" Dec 04 19:01:54 crc kubenswrapper[4733]: I1204 19:01:54.418549 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="fea0eafe-f021-49e1-b949-450a65b297db" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.245:5672: connect: connection refused" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.249350 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.265887 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-plugins\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.265971 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-confd\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.266128 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-erlang-cookie\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.266368 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.266417 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghm78\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-kube-api-access-ghm78\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.266444 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc170837-84b1-4112-b0fc-e483e9df9f60-erlang-cookie-secret\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.266470 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc170837-84b1-4112-b0fc-e483e9df9f60-pod-info\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.266493 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-server-conf\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.266522 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-plugins-conf\") pod \"bc170837-84b1-4112-b0fc-e483e9df9f60\" (UID: \"bc170837-84b1-4112-b0fc-e483e9df9f60\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.267595 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.268486 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.268821 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.273957 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/bc170837-84b1-4112-b0fc-e483e9df9f60-pod-info" (OuterVolumeSpecName: "pod-info") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.274060 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-kube-api-access-ghm78" (OuterVolumeSpecName: "kube-api-access-ghm78") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "kube-api-access-ghm78". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.274236 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc170837-84b1-4112-b0fc-e483e9df9f60-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.328543 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-server-conf" (OuterVolumeSpecName: "server-conf") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.332596 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d" (OuterVolumeSpecName: "persistence") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "pvc-e65898df-0ae6-4d48-a2ed-f321d450173d". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.367449 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") on node \"crc\" " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.367489 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghm78\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-kube-api-access-ghm78\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.367502 4733 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc170837-84b1-4112-b0fc-e483e9df9f60-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.367512 4733 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc170837-84b1-4112-b0fc-e483e9df9f60-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.367522 4733 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.367530 4733 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc170837-84b1-4112-b0fc-e483e9df9f60-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.367538 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.367546 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.397629 4733 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.397808 4733 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e65898df-0ae6-4d48-a2ed-f321d450173d" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d") on node "crc" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.410113 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "bc170837-84b1-4112-b0fc-e483e9df9f60" (UID: "bc170837-84b1-4112-b0fc-e483e9df9f60"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.442043 4733 generic.go:334] "Generic (PLEG): container finished" podID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerID="4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a" exitCode=0 Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.442335 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc170837-84b1-4112-b0fc-e483e9df9f60","Type":"ContainerDied","Data":"4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a"} Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.442385 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bc170837-84b1-4112-b0fc-e483e9df9f60","Type":"ContainerDied","Data":"ffb3459ab3e46a3142910e71e264cb94dfbae4edd7c2055488283afc2ef38712"} Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.442403 4733 scope.go:117] "RemoveContainer" containerID="4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.442593 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.468902 4733 reconciler_common.go:293] "Volume detached for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.468939 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc170837-84b1-4112-b0fc-e483e9df9f60-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.520172 4733 scope.go:117] "RemoveContainer" containerID="db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.535451 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.555055 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.562182 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:01:59 crc kubenswrapper[4733]: E1204 19:01:59.562539 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerName="rabbitmq" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.562559 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerName="rabbitmq" Dec 04 19:01:59 crc kubenswrapper[4733]: E1204 19:01:59.562577 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerName="setup-container" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.562585 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerName="setup-container" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.562737 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc170837-84b1-4112-b0fc-e483e9df9f60" containerName="rabbitmq" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.566731 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.572251 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.572630 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.573209 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-4hz4z" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.574518 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.574917 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.576045 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.630976 4733 scope.go:117] "RemoveContainer" containerID="4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a" Dec 04 19:01:59 crc kubenswrapper[4733]: E1204 19:01:59.631370 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a\": container with ID starting with 4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a not found: ID does not exist" containerID="4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.631401 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a"} err="failed to get container status \"4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a\": rpc error: code = NotFound desc = could not find container \"4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a\": container with ID starting with 4264127ac85658c65b068ce015c6373c6094746e49152787dda49836adfa375a not found: ID does not exist" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.631422 4733 scope.go:117] "RemoveContainer" containerID="db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e" Dec 04 19:01:59 crc kubenswrapper[4733]: E1204 19:01:59.631782 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e\": container with ID starting with db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e not found: ID does not exist" containerID="db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.631849 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e"} err="failed to get container status \"db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e\": rpc error: code = NotFound desc = could not find container \"db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e\": container with ID starting with db4800290de70cc9b6e9ab99f40fc3a4187c495d39d0c5d4fe6dd4d5751aa02e not found: ID does not exist" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675400 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675444 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9eb02384-3d2d-4977-9896-b57767861dbb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675508 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9eb02384-3d2d-4977-9896-b57767861dbb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675526 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9eb02384-3d2d-4977-9896-b57767861dbb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675608 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9eb02384-3d2d-4977-9896-b57767861dbb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675743 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675762 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.675785 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dcjd\" (UniqueName: \"kubernetes.io/projected/9eb02384-3d2d-4977-9896-b57767861dbb-kube-api-access-2dcjd\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.777634 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.777910 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9eb02384-3d2d-4977-9896-b57767861dbb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.777942 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.777974 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9eb02384-3d2d-4977-9896-b57767861dbb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.777992 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9eb02384-3d2d-4977-9896-b57767861dbb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.778012 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9eb02384-3d2d-4977-9896-b57767861dbb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.778040 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.778058 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.779003 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dcjd\" (UniqueName: \"kubernetes.io/projected/9eb02384-3d2d-4977-9896-b57767861dbb-kube-api-access-2dcjd\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.779313 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.779415 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9eb02384-3d2d-4977-9896-b57767861dbb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.779419 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9eb02384-3d2d-4977-9896-b57767861dbb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.779996 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.781200 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.781229 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bff92f9126deea5cf8426ea124f9832b721e3fe298f7a31ffdaeaf0c710d38b4/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.782696 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9eb02384-3d2d-4977-9896-b57767861dbb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.782856 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9eb02384-3d2d-4977-9896-b57767861dbb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.783959 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9eb02384-3d2d-4977-9896-b57767861dbb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.797347 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dcjd\" (UniqueName: \"kubernetes.io/projected/9eb02384-3d2d-4977-9896-b57767861dbb-kube-api-access-2dcjd\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.812231 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e65898df-0ae6-4d48-a2ed-f321d450173d\") pod \"rabbitmq-server-0\" (UID: \"9eb02384-3d2d-4977-9896-b57767861dbb\") " pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.827473 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.901552 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982159 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-confd\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982239 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-plugins-conf\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982262 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-plugins\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982298 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-erlang-cookie\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982348 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvmsd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-kube-api-access-tvmsd\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982368 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fea0eafe-f021-49e1-b949-450a65b297db-pod-info\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982406 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fea0eafe-f021-49e1-b949-450a65b297db-erlang-cookie-secret\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982507 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.982539 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-server-conf\") pod \"fea0eafe-f021-49e1-b949-450a65b297db\" (UID: \"fea0eafe-f021-49e1-b949-450a65b297db\") " Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.983562 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.983583 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.986119 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.987044 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/fea0eafe-f021-49e1-b949-450a65b297db-pod-info" (OuterVolumeSpecName: "pod-info") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.987131 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea0eafe-f021-49e1-b949-450a65b297db-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:01:59 crc kubenswrapper[4733]: I1204 19:01:59.995188 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d" (OuterVolumeSpecName: "persistence") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.004004 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-server-conf" (OuterVolumeSpecName: "server-conf") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.005009 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-kube-api-access-tvmsd" (OuterVolumeSpecName: "kube-api-access-tvmsd") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "kube-api-access-tvmsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.076399 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "fea0eafe-f021-49e1-b949-450a65b297db" (UID: "fea0eafe-f021-49e1-b949-450a65b297db"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084365 4733 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084403 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084419 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084433 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvmsd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-kube-api-access-tvmsd\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084445 4733 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fea0eafe-f021-49e1-b949-450a65b297db-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084457 4733 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fea0eafe-f021-49e1-b949-450a65b297db-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084505 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") on node \"crc\" " Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084521 4733 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fea0eafe-f021-49e1-b949-450a65b297db-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.084533 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fea0eafe-f021-49e1-b949-450a65b297db-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.104352 4733 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.104533 4733 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d") on node "crc" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.186098 4733 reconciler_common.go:293] "Volume detached for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.255981 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.351156 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc170837-84b1-4112-b0fc-e483e9df9f60" path="/var/lib/kubelet/pods/bc170837-84b1-4112-b0fc-e483e9df9f60/volumes" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.352496 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cbb4f659c-sbt7t"] Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.353275 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" podUID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" containerName="dnsmasq-dns" containerID="cri-o://56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233" gracePeriod=10 Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.400157 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 19:02:00 crc kubenswrapper[4733]: W1204 19:02:00.439110 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eb02384_3d2d_4977_9896_b57767861dbb.slice/crio-aedc772e510cd110ffd24016f39ac7db3674dc58047992e9f21c6a20ed4e94d1 WatchSource:0}: Error finding container aedc772e510cd110ffd24016f39ac7db3674dc58047992e9f21c6a20ed4e94d1: Status 404 returned error can't find the container with id aedc772e510cd110ffd24016f39ac7db3674dc58047992e9f21c6a20ed4e94d1 Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.453010 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9eb02384-3d2d-4977-9896-b57767861dbb","Type":"ContainerStarted","Data":"aedc772e510cd110ffd24016f39ac7db3674dc58047992e9f21c6a20ed4e94d1"} Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.454832 4733 generic.go:334] "Generic (PLEG): container finished" podID="fea0eafe-f021-49e1-b949-450a65b297db" containerID="c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e" exitCode=0 Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.454874 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fea0eafe-f021-49e1-b949-450a65b297db","Type":"ContainerDied","Data":"c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e"} Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.454891 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fea0eafe-f021-49e1-b949-450a65b297db","Type":"ContainerDied","Data":"625a075da08bebed457b7094108e5851b6fd71f8379af91c0d24052bec4ac93d"} Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.454906 4733 scope.go:117] "RemoveContainer" containerID="c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.454994 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.487987 4733 scope.go:117] "RemoveContainer" containerID="b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.495147 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.509746 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.519684 4733 scope.go:117] "RemoveContainer" containerID="c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.519787 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:02:00 crc kubenswrapper[4733]: E1204 19:02:00.520160 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea0eafe-f021-49e1-b949-450a65b297db" containerName="setup-container" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.520188 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea0eafe-f021-49e1-b949-450a65b297db" containerName="setup-container" Dec 04 19:02:00 crc kubenswrapper[4733]: E1204 19:02:00.520212 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea0eafe-f021-49e1-b949-450a65b297db" containerName="rabbitmq" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.520221 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea0eafe-f021-49e1-b949-450a65b297db" containerName="rabbitmq" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.520414 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="fea0eafe-f021-49e1-b949-450a65b297db" containerName="rabbitmq" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.521380 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.524427 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.524427 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.524514 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.524528 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.524876 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gpm72" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.530182 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:02:00 crc kubenswrapper[4733]: E1204 19:02:00.542398 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e\": container with ID starting with c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e not found: ID does not exist" containerID="c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.542483 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e"} err="failed to get container status \"c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e\": rpc error: code = NotFound desc = could not find container \"c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e\": container with ID starting with c4bc903835815332a4f808deb0d05f914e091048608ed069d4e9698938d6162e not found: ID does not exist" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.542543 4733 scope.go:117] "RemoveContainer" containerID="b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816" Dec 04 19:02:00 crc kubenswrapper[4733]: E1204 19:02:00.546931 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816\": container with ID starting with b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816 not found: ID does not exist" containerID="b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.546974 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816"} err="failed to get container status \"b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816\": rpc error: code = NotFound desc = could not find container \"b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816\": container with ID starting with b5fdd9a74099d9fb5158f36ef02e2a359cf363e3d1b4e102f2a31f2a58f25816 not found: ID does not exist" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.696405 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.696476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.696501 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.696639 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n27s\" (UniqueName: \"kubernetes.io/projected/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-kube-api-access-4n27s\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.696736 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.696766 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.696983 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.697042 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.697095 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.785028 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804138 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804190 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804219 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804247 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804287 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804348 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n27s\" (UniqueName: \"kubernetes.io/projected/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-kube-api-access-4n27s\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804371 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.804920 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.807236 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.807895 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.808630 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.813740 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.814189 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.822372 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.849544 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n27s\" (UniqueName: \"kubernetes.io/projected/f9f9453e-eacc-4174-b5d5-30ccd949ee2b-kube-api-access-4n27s\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.905766 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-dns-svc\") pod \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.905840 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-config\") pod \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.905879 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fr8z\" (UniqueName: \"kubernetes.io/projected/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-kube-api-access-4fr8z\") pod \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\" (UID: \"8ae9da60-b365-4932-87d6-6bc9fe29d2cb\") " Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.922343 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-kube-api-access-4fr8z" (OuterVolumeSpecName: "kube-api-access-4fr8z") pod "8ae9da60-b365-4932-87d6-6bc9fe29d2cb" (UID: "8ae9da60-b365-4932-87d6-6bc9fe29d2cb"). InnerVolumeSpecName "kube-api-access-4fr8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.928984 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.929068 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2a402ac0835941d16d1fb7a83ff5ff9713c626041a7644231b79bfaf8135d6ab/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.958557 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ae9da60-b365-4932-87d6-6bc9fe29d2cb" (UID: "8ae9da60-b365-4932-87d6-6bc9fe29d2cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.968600 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-config" (OuterVolumeSpecName: "config") pod "8ae9da60-b365-4932-87d6-6bc9fe29d2cb" (UID: "8ae9da60-b365-4932-87d6-6bc9fe29d2cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:02:00 crc kubenswrapper[4733]: I1204 19:02:00.989981 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2215f0fd-ffd6-475d-9d4f-3d2c4ca2158d\") pod \"rabbitmq-cell1-server-0\" (UID: \"f9f9453e-eacc-4174-b5d5-30ccd949ee2b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.007646 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.007900 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.007966 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fr8z\" (UniqueName: \"kubernetes.io/projected/8ae9da60-b365-4932-87d6-6bc9fe29d2cb-kube-api-access-4fr8z\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.157864 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.406393 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.469433 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" containerID="56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233" exitCode=0 Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.469531 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" event={"ID":"8ae9da60-b365-4932-87d6-6bc9fe29d2cb","Type":"ContainerDied","Data":"56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233"} Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.469598 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" event={"ID":"8ae9da60-b365-4932-87d6-6bc9fe29d2cb","Type":"ContainerDied","Data":"52f3d1c89f0bc68079cca98bf6370d6b6dd9e68f9fb252bf1c1e4c03a2ef1849"} Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.469624 4733 scope.go:117] "RemoveContainer" containerID="56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.470864 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cbb4f659c-sbt7t" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.473665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f9f9453e-eacc-4174-b5d5-30ccd949ee2b","Type":"ContainerStarted","Data":"9cf2c5159f68df03aa38a2aded88a9de5b1b6ae51937d17099f74157215e230d"} Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.512742 4733 scope.go:117] "RemoveContainer" containerID="2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.516743 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cbb4f659c-sbt7t"] Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.525978 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cbb4f659c-sbt7t"] Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.545674 4733 scope.go:117] "RemoveContainer" containerID="56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233" Dec 04 19:02:01 crc kubenswrapper[4733]: E1204 19:02:01.546403 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233\": container with ID starting with 56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233 not found: ID does not exist" containerID="56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.546457 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233"} err="failed to get container status \"56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233\": rpc error: code = NotFound desc = could not find container \"56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233\": container with ID starting with 56a3b06544495a2327302b8d97ed30f37c2c23cf6a8b6e31d94b67ac5a629233 not found: ID does not exist" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.546495 4733 scope.go:117] "RemoveContainer" containerID="2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b" Dec 04 19:02:01 crc kubenswrapper[4733]: E1204 19:02:01.547538 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b\": container with ID starting with 2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b not found: ID does not exist" containerID="2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b" Dec 04 19:02:01 crc kubenswrapper[4733]: I1204 19:02:01.547586 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b"} err="failed to get container status \"2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b\": rpc error: code = NotFound desc = could not find container \"2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b\": container with ID starting with 2003d565c136f4186139722187a86c15a4b59b6f2d2acaa7073b2e047a440c6b not found: ID does not exist" Dec 04 19:02:02 crc kubenswrapper[4733]: I1204 19:02:02.353528 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" path="/var/lib/kubelet/pods/8ae9da60-b365-4932-87d6-6bc9fe29d2cb/volumes" Dec 04 19:02:02 crc kubenswrapper[4733]: I1204 19:02:02.355960 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fea0eafe-f021-49e1-b949-450a65b297db" path="/var/lib/kubelet/pods/fea0eafe-f021-49e1-b949-450a65b297db/volumes" Dec 04 19:02:02 crc kubenswrapper[4733]: I1204 19:02:02.487563 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9eb02384-3d2d-4977-9896-b57767861dbb","Type":"ContainerStarted","Data":"27b21c0c177a9b7562d327d105ca4cc1b68cd999e6a3ba01ba11cb4a65997519"} Dec 04 19:02:03 crc kubenswrapper[4733]: I1204 19:02:03.501470 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f9f9453e-eacc-4174-b5d5-30ccd949ee2b","Type":"ContainerStarted","Data":"2eb3c9e2859d2a111e6fc21c999f9593b58ee604d846653821cbd9bdc31f9c5a"} Dec 04 19:02:34 crc kubenswrapper[4733]: I1204 19:02:34.821489 4733 generic.go:334] "Generic (PLEG): container finished" podID="9eb02384-3d2d-4977-9896-b57767861dbb" containerID="27b21c0c177a9b7562d327d105ca4cc1b68cd999e6a3ba01ba11cb4a65997519" exitCode=0 Dec 04 19:02:34 crc kubenswrapper[4733]: I1204 19:02:34.821608 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9eb02384-3d2d-4977-9896-b57767861dbb","Type":"ContainerDied","Data":"27b21c0c177a9b7562d327d105ca4cc1b68cd999e6a3ba01ba11cb4a65997519"} Dec 04 19:02:35 crc kubenswrapper[4733]: I1204 19:02:35.832492 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9eb02384-3d2d-4977-9896-b57767861dbb","Type":"ContainerStarted","Data":"1cb999eea1464fa256cfff483f43d0e1ee6edae84ea6c66cfae0007c8d121e30"} Dec 04 19:02:35 crc kubenswrapper[4733]: I1204 19:02:35.832970 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 19:02:35 crc kubenswrapper[4733]: I1204 19:02:35.834359 4733 generic.go:334] "Generic (PLEG): container finished" podID="f9f9453e-eacc-4174-b5d5-30ccd949ee2b" containerID="2eb3c9e2859d2a111e6fc21c999f9593b58ee604d846653821cbd9bdc31f9c5a" exitCode=0 Dec 04 19:02:35 crc kubenswrapper[4733]: I1204 19:02:35.834406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f9f9453e-eacc-4174-b5d5-30ccd949ee2b","Type":"ContainerDied","Data":"2eb3c9e2859d2a111e6fc21c999f9593b58ee604d846653821cbd9bdc31f9c5a"} Dec 04 19:02:35 crc kubenswrapper[4733]: I1204 19:02:35.861214 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.861192763 podStartE2EDuration="36.861192763s" podCreationTimestamp="2025-12-04 19:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:02:35.856171987 +0000 UTC m=+5017.811533043" watchObservedRunningTime="2025-12-04 19:02:35.861192763 +0000 UTC m=+5017.816553809" Dec 04 19:02:36 crc kubenswrapper[4733]: I1204 19:02:36.845174 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f9f9453e-eacc-4174-b5d5-30ccd949ee2b","Type":"ContainerStarted","Data":"2284e9139c81ff3f529dc84f8f89b8be31d156f219648f5d8e9e2c975c4cedaf"} Dec 04 19:02:36 crc kubenswrapper[4733]: I1204 19:02:36.845804 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:36 crc kubenswrapper[4733]: I1204 19:02:36.881168 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.881154506 podStartE2EDuration="36.881154506s" podCreationTimestamp="2025-12-04 19:02:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:02:36.877749194 +0000 UTC m=+5018.833110260" watchObservedRunningTime="2025-12-04 19:02:36.881154506 +0000 UTC m=+5018.836515552" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.470399 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nmtl6"] Dec 04 19:02:44 crc kubenswrapper[4733]: E1204 19:02:44.471329 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" containerName="dnsmasq-dns" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.471346 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" containerName="dnsmasq-dns" Dec 04 19:02:44 crc kubenswrapper[4733]: E1204 19:02:44.471372 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" containerName="init" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.471378 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" containerName="init" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.471568 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae9da60-b365-4932-87d6-6bc9fe29d2cb" containerName="dnsmasq-dns" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.472904 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.517723 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-utilities\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.517801 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-catalog-content\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.517830 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m4wq\" (UniqueName: \"kubernetes.io/projected/17bc0ac7-83b2-47e1-896f-39b89a600f93-kube-api-access-2m4wq\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.529936 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmtl6"] Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.618986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-catalog-content\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.619048 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m4wq\" (UniqueName: \"kubernetes.io/projected/17bc0ac7-83b2-47e1-896f-39b89a600f93-kube-api-access-2m4wq\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.619152 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-utilities\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.619471 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-catalog-content\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.619619 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-utilities\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.639301 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m4wq\" (UniqueName: \"kubernetes.io/projected/17bc0ac7-83b2-47e1-896f-39b89a600f93-kube-api-access-2m4wq\") pod \"redhat-marketplace-nmtl6\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:44 crc kubenswrapper[4733]: I1204 19:02:44.833872 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:45 crc kubenswrapper[4733]: I1204 19:02:45.362340 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:02:45 crc kubenswrapper[4733]: I1204 19:02:45.362647 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:02:45 crc kubenswrapper[4733]: I1204 19:02:45.845486 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmtl6"] Dec 04 19:02:45 crc kubenswrapper[4733]: W1204 19:02:45.856292 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17bc0ac7_83b2_47e1_896f_39b89a600f93.slice/crio-dc434b44f0de970f5915ba3d2932b2201d290c3e30ce30305ed8ab8686cc66d6 WatchSource:0}: Error finding container dc434b44f0de970f5915ba3d2932b2201d290c3e30ce30305ed8ab8686cc66d6: Status 404 returned error can't find the container with id dc434b44f0de970f5915ba3d2932b2201d290c3e30ce30305ed8ab8686cc66d6 Dec 04 19:02:45 crc kubenswrapper[4733]: I1204 19:02:45.926451 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmtl6" event={"ID":"17bc0ac7-83b2-47e1-896f-39b89a600f93","Type":"ContainerStarted","Data":"dc434b44f0de970f5915ba3d2932b2201d290c3e30ce30305ed8ab8686cc66d6"} Dec 04 19:02:46 crc kubenswrapper[4733]: I1204 19:02:46.937062 4733 generic.go:334] "Generic (PLEG): container finished" podID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerID="430eecf1f6573ad992ee7c3b17e2ab6af39b0e71ff55d2c3a55cbc16d483fd30" exitCode=0 Dec 04 19:02:46 crc kubenswrapper[4733]: I1204 19:02:46.937126 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmtl6" event={"ID":"17bc0ac7-83b2-47e1-896f-39b89a600f93","Type":"ContainerDied","Data":"430eecf1f6573ad992ee7c3b17e2ab6af39b0e71ff55d2c3a55cbc16d483fd30"} Dec 04 19:02:46 crc kubenswrapper[4733]: I1204 19:02:46.941101 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:02:47 crc kubenswrapper[4733]: I1204 19:02:47.948975 4733 generic.go:334] "Generic (PLEG): container finished" podID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerID="3c41b451e4b490edb182188de742cb0019132e410db926c309e2c1c94a633552" exitCode=0 Dec 04 19:02:47 crc kubenswrapper[4733]: I1204 19:02:47.949050 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmtl6" event={"ID":"17bc0ac7-83b2-47e1-896f-39b89a600f93","Type":"ContainerDied","Data":"3c41b451e4b490edb182188de742cb0019132e410db926c309e2c1c94a633552"} Dec 04 19:02:48 crc kubenswrapper[4733]: I1204 19:02:48.961907 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmtl6" event={"ID":"17bc0ac7-83b2-47e1-896f-39b89a600f93","Type":"ContainerStarted","Data":"8562146b562bce7b7de00cc9488cc6cff9cbaf5997b4d0ac78a6f884b0997ea2"} Dec 04 19:02:49 crc kubenswrapper[4733]: I1204 19:02:49.904971 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 19:02:49 crc kubenswrapper[4733]: I1204 19:02:49.926663 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nmtl6" podStartSLOduration=4.493844258 podStartE2EDuration="5.926641388s" podCreationTimestamp="2025-12-04 19:02:44 +0000 UTC" firstStartedPulling="2025-12-04 19:02:46.940726536 +0000 UTC m=+5028.896087602" lastFinishedPulling="2025-12-04 19:02:48.373523686 +0000 UTC m=+5030.328884732" observedRunningTime="2025-12-04 19:02:49.00529497 +0000 UTC m=+5030.960656016" watchObservedRunningTime="2025-12-04 19:02:49.926641388 +0000 UTC m=+5031.882002434" Dec 04 19:02:51 crc kubenswrapper[4733]: I1204 19:02:51.162010 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 19:02:54 crc kubenswrapper[4733]: I1204 19:02:54.834137 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:54 crc kubenswrapper[4733]: I1204 19:02:54.834458 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:54 crc kubenswrapper[4733]: I1204 19:02:54.901832 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:55 crc kubenswrapper[4733]: I1204 19:02:55.057050 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:55 crc kubenswrapper[4733]: I1204 19:02:55.149896 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmtl6"] Dec 04 19:02:57 crc kubenswrapper[4733]: I1204 19:02:57.030919 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nmtl6" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerName="registry-server" containerID="cri-o://8562146b562bce7b7de00cc9488cc6cff9cbaf5997b4d0ac78a6f884b0997ea2" gracePeriod=2 Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.039384 4733 generic.go:334] "Generic (PLEG): container finished" podID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerID="8562146b562bce7b7de00cc9488cc6cff9cbaf5997b4d0ac78a6f884b0997ea2" exitCode=0 Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.039458 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmtl6" event={"ID":"17bc0ac7-83b2-47e1-896f-39b89a600f93","Type":"ContainerDied","Data":"8562146b562bce7b7de00cc9488cc6cff9cbaf5997b4d0ac78a6f884b0997ea2"} Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.039783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmtl6" event={"ID":"17bc0ac7-83b2-47e1-896f-39b89a600f93","Type":"ContainerDied","Data":"dc434b44f0de970f5915ba3d2932b2201d290c3e30ce30305ed8ab8686cc66d6"} Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.039812 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc434b44f0de970f5915ba3d2932b2201d290c3e30ce30305ed8ab8686cc66d6" Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.072407 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.232279 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-utilities\") pod \"17bc0ac7-83b2-47e1-896f-39b89a600f93\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.232421 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-catalog-content\") pod \"17bc0ac7-83b2-47e1-896f-39b89a600f93\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.232555 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m4wq\" (UniqueName: \"kubernetes.io/projected/17bc0ac7-83b2-47e1-896f-39b89a600f93-kube-api-access-2m4wq\") pod \"17bc0ac7-83b2-47e1-896f-39b89a600f93\" (UID: \"17bc0ac7-83b2-47e1-896f-39b89a600f93\") " Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.233376 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-utilities" (OuterVolumeSpecName: "utilities") pod "17bc0ac7-83b2-47e1-896f-39b89a600f93" (UID: "17bc0ac7-83b2-47e1-896f-39b89a600f93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.240879 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17bc0ac7-83b2-47e1-896f-39b89a600f93-kube-api-access-2m4wq" (OuterVolumeSpecName: "kube-api-access-2m4wq") pod "17bc0ac7-83b2-47e1-896f-39b89a600f93" (UID: "17bc0ac7-83b2-47e1-896f-39b89a600f93"). InnerVolumeSpecName "kube-api-access-2m4wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.254148 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17bc0ac7-83b2-47e1-896f-39b89a600f93" (UID: "17bc0ac7-83b2-47e1-896f-39b89a600f93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.335160 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.335631 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17bc0ac7-83b2-47e1-896f-39b89a600f93-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:58 crc kubenswrapper[4733]: I1204 19:02:58.335706 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m4wq\" (UniqueName: \"kubernetes.io/projected/17bc0ac7-83b2-47e1-896f-39b89a600f93-kube-api-access-2m4wq\") on node \"crc\" DevicePath \"\"" Dec 04 19:02:59 crc kubenswrapper[4733]: I1204 19:02:59.047513 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nmtl6" Dec 04 19:02:59 crc kubenswrapper[4733]: I1204 19:02:59.075863 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmtl6"] Dec 04 19:02:59 crc kubenswrapper[4733]: I1204 19:02:59.084182 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmtl6"] Dec 04 19:03:00 crc kubenswrapper[4733]: I1204 19:03:00.353743 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" path="/var/lib/kubelet/pods/17bc0ac7-83b2-47e1-896f-39b89a600f93/volumes" Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.728480 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 19:03:02 crc kubenswrapper[4733]: E1204 19:03:02.729612 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerName="registry-server" Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.729702 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerName="registry-server" Dec 04 19:03:02 crc kubenswrapper[4733]: E1204 19:03:02.729760 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerName="extract-content" Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.729830 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerName="extract-content" Dec 04 19:03:02 crc kubenswrapper[4733]: E1204 19:03:02.729884 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerName="extract-utilities" Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.729926 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerName="extract-utilities" Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.730119 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="17bc0ac7-83b2-47e1-896f-39b89a600f93" containerName="registry-server" Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.730776 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.733539 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-hvvvj" Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.743452 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 19:03:02 crc kubenswrapper[4733]: I1204 19:03:02.909559 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg6lm\" (UniqueName: \"kubernetes.io/projected/ce858752-a366-436b-a574-a8b2582282b8-kube-api-access-mg6lm\") pod \"mariadb-client-1-default\" (UID: \"ce858752-a366-436b-a574-a8b2582282b8\") " pod="openstack/mariadb-client-1-default" Dec 04 19:03:03 crc kubenswrapper[4733]: I1204 19:03:03.011439 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg6lm\" (UniqueName: \"kubernetes.io/projected/ce858752-a366-436b-a574-a8b2582282b8-kube-api-access-mg6lm\") pod \"mariadb-client-1-default\" (UID: \"ce858752-a366-436b-a574-a8b2582282b8\") " pod="openstack/mariadb-client-1-default" Dec 04 19:03:03 crc kubenswrapper[4733]: I1204 19:03:03.043714 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg6lm\" (UniqueName: \"kubernetes.io/projected/ce858752-a366-436b-a574-a8b2582282b8-kube-api-access-mg6lm\") pod \"mariadb-client-1-default\" (UID: \"ce858752-a366-436b-a574-a8b2582282b8\") " pod="openstack/mariadb-client-1-default" Dec 04 19:03:03 crc kubenswrapper[4733]: I1204 19:03:03.050421 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 04 19:03:03 crc kubenswrapper[4733]: I1204 19:03:03.403916 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 19:03:03 crc kubenswrapper[4733]: W1204 19:03:03.414560 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce858752_a366_436b_a574_a8b2582282b8.slice/crio-fb2af58a45c427fc31d90d0ce6b941228f5fb404e02f08f9e4c47d5107847a2a WatchSource:0}: Error finding container fb2af58a45c427fc31d90d0ce6b941228f5fb404e02f08f9e4c47d5107847a2a: Status 404 returned error can't find the container with id fb2af58a45c427fc31d90d0ce6b941228f5fb404e02f08f9e4c47d5107847a2a Dec 04 19:03:04 crc kubenswrapper[4733]: I1204 19:03:04.093901 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"ce858752-a366-436b-a574-a8b2582282b8","Type":"ContainerStarted","Data":"fb2af58a45c427fc31d90d0ce6b941228f5fb404e02f08f9e4c47d5107847a2a"} Dec 04 19:03:05 crc kubenswrapper[4733]: I1204 19:03:05.105448 4733 generic.go:334] "Generic (PLEG): container finished" podID="ce858752-a366-436b-a574-a8b2582282b8" containerID="8e9944637059f44dd22dd9c6dddac0be3c602e2776f7be8c448be57e4255534e" exitCode=0 Dec 04 19:03:05 crc kubenswrapper[4733]: I1204 19:03:05.105522 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"ce858752-a366-436b-a574-a8b2582282b8","Type":"ContainerDied","Data":"8e9944637059f44dd22dd9c6dddac0be3c602e2776f7be8c448be57e4255534e"} Dec 04 19:03:06 crc kubenswrapper[4733]: I1204 19:03:06.538555 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 04 19:03:06 crc kubenswrapper[4733]: I1204 19:03:06.574218 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_ce858752-a366-436b-a574-a8b2582282b8/mariadb-client-1-default/0.log" Dec 04 19:03:06 crc kubenswrapper[4733]: I1204 19:03:06.623827 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 19:03:06 crc kubenswrapper[4733]: I1204 19:03:06.633140 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 19:03:06 crc kubenswrapper[4733]: I1204 19:03:06.665964 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg6lm\" (UniqueName: \"kubernetes.io/projected/ce858752-a366-436b-a574-a8b2582282b8-kube-api-access-mg6lm\") pod \"ce858752-a366-436b-a574-a8b2582282b8\" (UID: \"ce858752-a366-436b-a574-a8b2582282b8\") " Dec 04 19:03:06 crc kubenswrapper[4733]: I1204 19:03:06.692579 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce858752-a366-436b-a574-a8b2582282b8-kube-api-access-mg6lm" (OuterVolumeSpecName: "kube-api-access-mg6lm") pod "ce858752-a366-436b-a574-a8b2582282b8" (UID: "ce858752-a366-436b-a574-a8b2582282b8"). InnerVolumeSpecName "kube-api-access-mg6lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:03:06 crc kubenswrapper[4733]: I1204 19:03:06.767361 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg6lm\" (UniqueName: \"kubernetes.io/projected/ce858752-a366-436b-a574-a8b2582282b8-kube-api-access-mg6lm\") on node \"crc\" DevicePath \"\"" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.056293 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 19:03:07 crc kubenswrapper[4733]: E1204 19:03:07.056594 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce858752-a366-436b-a574-a8b2582282b8" containerName="mariadb-client-1-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.056611 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce858752-a366-436b-a574-a8b2582282b8" containerName="mariadb-client-1-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.056772 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce858752-a366-436b-a574-a8b2582282b8" containerName="mariadb-client-1-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.057353 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.063214 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.110331 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w7x2\" (UniqueName: \"kubernetes.io/projected/a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d-kube-api-access-8w7x2\") pod \"mariadb-client-2-default\" (UID: \"a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d\") " pod="openstack/mariadb-client-2-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.148785 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb2af58a45c427fc31d90d0ce6b941228f5fb404e02f08f9e4c47d5107847a2a" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.148958 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.211583 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w7x2\" (UniqueName: \"kubernetes.io/projected/a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d-kube-api-access-8w7x2\") pod \"mariadb-client-2-default\" (UID: \"a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d\") " pod="openstack/mariadb-client-2-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.227592 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w7x2\" (UniqueName: \"kubernetes.io/projected/a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d-kube-api-access-8w7x2\") pod \"mariadb-client-2-default\" (UID: \"a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d\") " pod="openstack/mariadb-client-2-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.447060 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 04 19:03:07 crc kubenswrapper[4733]: I1204 19:03:07.986141 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 19:03:07 crc kubenswrapper[4733]: W1204 19:03:07.998706 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda73adc61_bf02_4fd3_a2fc_8cb7e70bf80d.slice/crio-64867dcb860fd35b51537c7c51df1d0b8d8c3395b81ea5bede0fc3a48266e5cf WatchSource:0}: Error finding container 64867dcb860fd35b51537c7c51df1d0b8d8c3395b81ea5bede0fc3a48266e5cf: Status 404 returned error can't find the container with id 64867dcb860fd35b51537c7c51df1d0b8d8c3395b81ea5bede0fc3a48266e5cf Dec 04 19:03:08 crc kubenswrapper[4733]: I1204 19:03:08.157652 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d","Type":"ContainerStarted","Data":"64867dcb860fd35b51537c7c51df1d0b8d8c3395b81ea5bede0fc3a48266e5cf"} Dec 04 19:03:08 crc kubenswrapper[4733]: I1204 19:03:08.345227 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce858752-a366-436b-a574-a8b2582282b8" path="/var/lib/kubelet/pods/ce858752-a366-436b-a574-a8b2582282b8/volumes" Dec 04 19:03:09 crc kubenswrapper[4733]: I1204 19:03:09.166686 4733 generic.go:334] "Generic (PLEG): container finished" podID="a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d" containerID="e8a595f3dde8366c3cad95b7ae6cd3edc07217fb3dbcf1d87a183f1f0b58fdcc" exitCode=1 Dec 04 19:03:09 crc kubenswrapper[4733]: I1204 19:03:09.166730 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d","Type":"ContainerDied","Data":"e8a595f3dde8366c3cad95b7ae6cd3edc07217fb3dbcf1d87a183f1f0b58fdcc"} Dec 04 19:03:10 crc kubenswrapper[4733]: I1204 19:03:10.615835 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 04 19:03:10 crc kubenswrapper[4733]: I1204 19:03:10.632197 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d/mariadb-client-2-default/0.log" Dec 04 19:03:10 crc kubenswrapper[4733]: I1204 19:03:10.661406 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 19:03:10 crc kubenswrapper[4733]: I1204 19:03:10.666572 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w7x2\" (UniqueName: \"kubernetes.io/projected/a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d-kube-api-access-8w7x2\") pod \"a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d\" (UID: \"a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d\") " Dec 04 19:03:10 crc kubenswrapper[4733]: I1204 19:03:10.670060 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 19:03:10 crc kubenswrapper[4733]: I1204 19:03:10.673077 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d-kube-api-access-8w7x2" (OuterVolumeSpecName: "kube-api-access-8w7x2") pod "a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d" (UID: "a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d"). InnerVolumeSpecName "kube-api-access-8w7x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:03:10 crc kubenswrapper[4733]: I1204 19:03:10.769406 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w7x2\" (UniqueName: \"kubernetes.io/projected/a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d-kube-api-access-8w7x2\") on node \"crc\" DevicePath \"\"" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.051353 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Dec 04 19:03:11 crc kubenswrapper[4733]: E1204 19:03:11.051642 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d" containerName="mariadb-client-2-default" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.051660 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d" containerName="mariadb-client-2-default" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.051845 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d" containerName="mariadb-client-2-default" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.052296 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.061976 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.175353 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk927\" (UniqueName: \"kubernetes.io/projected/d58bc4b9-967a-46d7-a306-0ec6544cb567-kube-api-access-xk927\") pod \"mariadb-client-1\" (UID: \"d58bc4b9-967a-46d7-a306-0ec6544cb567\") " pod="openstack/mariadb-client-1" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.186140 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64867dcb860fd35b51537c7c51df1d0b8d8c3395b81ea5bede0fc3a48266e5cf" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.186211 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.276399 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk927\" (UniqueName: \"kubernetes.io/projected/d58bc4b9-967a-46d7-a306-0ec6544cb567-kube-api-access-xk927\") pod \"mariadb-client-1\" (UID: \"d58bc4b9-967a-46d7-a306-0ec6544cb567\") " pod="openstack/mariadb-client-1" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.303492 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk927\" (UniqueName: \"kubernetes.io/projected/d58bc4b9-967a-46d7-a306-0ec6544cb567-kube-api-access-xk927\") pod \"mariadb-client-1\" (UID: \"d58bc4b9-967a-46d7-a306-0ec6544cb567\") " pod="openstack/mariadb-client-1" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.379493 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 04 19:03:11 crc kubenswrapper[4733]: I1204 19:03:11.676080 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 04 19:03:12 crc kubenswrapper[4733]: I1204 19:03:12.198912 4733 generic.go:334] "Generic (PLEG): container finished" podID="d58bc4b9-967a-46d7-a306-0ec6544cb567" containerID="4e7afb3400424199209584a0266d44aa3af1590be06717fffc2b54a8d8dd38a1" exitCode=0 Dec 04 19:03:12 crc kubenswrapper[4733]: I1204 19:03:12.199042 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"d58bc4b9-967a-46d7-a306-0ec6544cb567","Type":"ContainerDied","Data":"4e7afb3400424199209584a0266d44aa3af1590be06717fffc2b54a8d8dd38a1"} Dec 04 19:03:12 crc kubenswrapper[4733]: I1204 19:03:12.199364 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"d58bc4b9-967a-46d7-a306-0ec6544cb567","Type":"ContainerStarted","Data":"a087a057c960af66306fe9066e050c42068c6be43e412c8531ea55222006bd23"} Dec 04 19:03:12 crc kubenswrapper[4733]: I1204 19:03:12.350672 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d" path="/var/lib/kubelet/pods/a73adc61-bf02-4fd3-a2fc-8cb7e70bf80d/volumes" Dec 04 19:03:13 crc kubenswrapper[4733]: I1204 19:03:13.647902 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 04 19:03:13 crc kubenswrapper[4733]: I1204 19:03:13.666109 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_d58bc4b9-967a-46d7-a306-0ec6544cb567/mariadb-client-1/0.log" Dec 04 19:03:13 crc kubenswrapper[4733]: I1204 19:03:13.698542 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Dec 04 19:03:13 crc kubenswrapper[4733]: I1204 19:03:13.704687 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Dec 04 19:03:13 crc kubenswrapper[4733]: I1204 19:03:13.817745 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk927\" (UniqueName: \"kubernetes.io/projected/d58bc4b9-967a-46d7-a306-0ec6544cb567-kube-api-access-xk927\") pod \"d58bc4b9-967a-46d7-a306-0ec6544cb567\" (UID: \"d58bc4b9-967a-46d7-a306-0ec6544cb567\") " Dec 04 19:03:13 crc kubenswrapper[4733]: I1204 19:03:13.825579 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d58bc4b9-967a-46d7-a306-0ec6544cb567-kube-api-access-xk927" (OuterVolumeSpecName: "kube-api-access-xk927") pod "d58bc4b9-967a-46d7-a306-0ec6544cb567" (UID: "d58bc4b9-967a-46d7-a306-0ec6544cb567"). InnerVolumeSpecName "kube-api-access-xk927". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:03:13 crc kubenswrapper[4733]: I1204 19:03:13.919993 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk927\" (UniqueName: \"kubernetes.io/projected/d58bc4b9-967a-46d7-a306-0ec6544cb567-kube-api-access-xk927\") on node \"crc\" DevicePath \"\"" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.125040 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 19:03:14 crc kubenswrapper[4733]: E1204 19:03:14.125651 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d58bc4b9-967a-46d7-a306-0ec6544cb567" containerName="mariadb-client-1" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.125736 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d58bc4b9-967a-46d7-a306-0ec6544cb567" containerName="mariadb-client-1" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.126005 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d58bc4b9-967a-46d7-a306-0ec6544cb567" containerName="mariadb-client-1" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.126736 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.156043 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.222339 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a087a057c960af66306fe9066e050c42068c6be43e412c8531ea55222006bd23" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.222401 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.226163 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2ldt\" (UniqueName: \"kubernetes.io/projected/61319776-682c-493b-87cf-920a9690d86b-kube-api-access-r2ldt\") pod \"mariadb-client-4-default\" (UID: \"61319776-682c-493b-87cf-920a9690d86b\") " pod="openstack/mariadb-client-4-default" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.327417 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2ldt\" (UniqueName: \"kubernetes.io/projected/61319776-682c-493b-87cf-920a9690d86b-kube-api-access-r2ldt\") pod \"mariadb-client-4-default\" (UID: \"61319776-682c-493b-87cf-920a9690d86b\") " pod="openstack/mariadb-client-4-default" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.346339 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d58bc4b9-967a-46d7-a306-0ec6544cb567" path="/var/lib/kubelet/pods/d58bc4b9-967a-46d7-a306-0ec6544cb567/volumes" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.347306 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2ldt\" (UniqueName: \"kubernetes.io/projected/61319776-682c-493b-87cf-920a9690d86b-kube-api-access-r2ldt\") pod \"mariadb-client-4-default\" (UID: \"61319776-682c-493b-87cf-920a9690d86b\") " pod="openstack/mariadb-client-4-default" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.446714 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 04 19:03:14 crc kubenswrapper[4733]: I1204 19:03:14.990106 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 19:03:14 crc kubenswrapper[4733]: W1204 19:03:14.992685 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61319776_682c_493b_87cf_920a9690d86b.slice/crio-fe10792260ac81f2c3ed60df16067ce62a8b6b39dfd9abbbdfeceb9a7cf99f14 WatchSource:0}: Error finding container fe10792260ac81f2c3ed60df16067ce62a8b6b39dfd9abbbdfeceb9a7cf99f14: Status 404 returned error can't find the container with id fe10792260ac81f2c3ed60df16067ce62a8b6b39dfd9abbbdfeceb9a7cf99f14 Dec 04 19:03:15 crc kubenswrapper[4733]: I1204 19:03:15.230398 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"61319776-682c-493b-87cf-920a9690d86b","Type":"ContainerStarted","Data":"fe10792260ac81f2c3ed60df16067ce62a8b6b39dfd9abbbdfeceb9a7cf99f14"} Dec 04 19:03:15 crc kubenswrapper[4733]: I1204 19:03:15.361621 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:03:15 crc kubenswrapper[4733]: I1204 19:03:15.361698 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:03:16 crc kubenswrapper[4733]: I1204 19:03:16.243205 4733 generic.go:334] "Generic (PLEG): container finished" podID="61319776-682c-493b-87cf-920a9690d86b" containerID="135aa5524cf9b40263a7992c7090b76cb026c548da8822eb31061e3ed969925c" exitCode=0 Dec 04 19:03:16 crc kubenswrapper[4733]: I1204 19:03:16.243407 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"61319776-682c-493b-87cf-920a9690d86b","Type":"ContainerDied","Data":"135aa5524cf9b40263a7992c7090b76cb026c548da8822eb31061e3ed969925c"} Dec 04 19:03:17 crc kubenswrapper[4733]: I1204 19:03:17.625817 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 04 19:03:17 crc kubenswrapper[4733]: I1204 19:03:17.647939 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_61319776-682c-493b-87cf-920a9690d86b/mariadb-client-4-default/0.log" Dec 04 19:03:17 crc kubenswrapper[4733]: I1204 19:03:17.673585 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 19:03:17 crc kubenswrapper[4733]: I1204 19:03:17.679081 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 19:03:17 crc kubenswrapper[4733]: I1204 19:03:17.783196 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2ldt\" (UniqueName: \"kubernetes.io/projected/61319776-682c-493b-87cf-920a9690d86b-kube-api-access-r2ldt\") pod \"61319776-682c-493b-87cf-920a9690d86b\" (UID: \"61319776-682c-493b-87cf-920a9690d86b\") " Dec 04 19:03:17 crc kubenswrapper[4733]: I1204 19:03:17.791381 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61319776-682c-493b-87cf-920a9690d86b-kube-api-access-r2ldt" (OuterVolumeSpecName: "kube-api-access-r2ldt") pod "61319776-682c-493b-87cf-920a9690d86b" (UID: "61319776-682c-493b-87cf-920a9690d86b"). InnerVolumeSpecName "kube-api-access-r2ldt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:03:17 crc kubenswrapper[4733]: I1204 19:03:17.885265 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2ldt\" (UniqueName: \"kubernetes.io/projected/61319776-682c-493b-87cf-920a9690d86b-kube-api-access-r2ldt\") on node \"crc\" DevicePath \"\"" Dec 04 19:03:18 crc kubenswrapper[4733]: I1204 19:03:18.261364 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe10792260ac81f2c3ed60df16067ce62a8b6b39dfd9abbbdfeceb9a7cf99f14" Dec 04 19:03:18 crc kubenswrapper[4733]: I1204 19:03:18.261467 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 04 19:03:18 crc kubenswrapper[4733]: I1204 19:03:18.352001 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61319776-682c-493b-87cf-920a9690d86b" path="/var/lib/kubelet/pods/61319776-682c-493b-87cf-920a9690d86b/volumes" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.206484 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 19:03:21 crc kubenswrapper[4733]: E1204 19:03:21.215021 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61319776-682c-493b-87cf-920a9690d86b" containerName="mariadb-client-4-default" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.215056 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="61319776-682c-493b-87cf-920a9690d86b" containerName="mariadb-client-4-default" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.215472 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="61319776-682c-493b-87cf-920a9690d86b" containerName="mariadb-client-4-default" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.216088 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.216172 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.223660 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-hvvvj" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.340939 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmgjr\" (UniqueName: \"kubernetes.io/projected/177fc63c-9a4e-4647-8b3f-d79137d096eb-kube-api-access-wmgjr\") pod \"mariadb-client-5-default\" (UID: \"177fc63c-9a4e-4647-8b3f-d79137d096eb\") " pod="openstack/mariadb-client-5-default" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.442913 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmgjr\" (UniqueName: \"kubernetes.io/projected/177fc63c-9a4e-4647-8b3f-d79137d096eb-kube-api-access-wmgjr\") pod \"mariadb-client-5-default\" (UID: \"177fc63c-9a4e-4647-8b3f-d79137d096eb\") " pod="openstack/mariadb-client-5-default" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.467887 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmgjr\" (UniqueName: \"kubernetes.io/projected/177fc63c-9a4e-4647-8b3f-d79137d096eb-kube-api-access-wmgjr\") pod \"mariadb-client-5-default\" (UID: \"177fc63c-9a4e-4647-8b3f-d79137d096eb\") " pod="openstack/mariadb-client-5-default" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.540553 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 04 19:03:21 crc kubenswrapper[4733]: I1204 19:03:21.872854 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 19:03:22 crc kubenswrapper[4733]: I1204 19:03:22.296125 4733 generic.go:334] "Generic (PLEG): container finished" podID="177fc63c-9a4e-4647-8b3f-d79137d096eb" containerID="ebfbed019bbe9565def43557bc96698a4c8ab45cf1c61fa9badc9e3ae490f7d8" exitCode=0 Dec 04 19:03:22 crc kubenswrapper[4733]: I1204 19:03:22.296215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"177fc63c-9a4e-4647-8b3f-d79137d096eb","Type":"ContainerDied","Data":"ebfbed019bbe9565def43557bc96698a4c8ab45cf1c61fa9badc9e3ae490f7d8"} Dec 04 19:03:22 crc kubenswrapper[4733]: I1204 19:03:22.296481 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"177fc63c-9a4e-4647-8b3f-d79137d096eb","Type":"ContainerStarted","Data":"21ac45a79b5027bd05a91a327f2ec2405d5d1541f48b9aa8c8fe27092a8c0567"} Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.649778 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.665477 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_177fc63c-9a4e-4647-8b3f-d79137d096eb/mariadb-client-5-default/0.log" Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.698961 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.705694 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.779468 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmgjr\" (UniqueName: \"kubernetes.io/projected/177fc63c-9a4e-4647-8b3f-d79137d096eb-kube-api-access-wmgjr\") pod \"177fc63c-9a4e-4647-8b3f-d79137d096eb\" (UID: \"177fc63c-9a4e-4647-8b3f-d79137d096eb\") " Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.787610 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/177fc63c-9a4e-4647-8b3f-d79137d096eb-kube-api-access-wmgjr" (OuterVolumeSpecName: "kube-api-access-wmgjr") pod "177fc63c-9a4e-4647-8b3f-d79137d096eb" (UID: "177fc63c-9a4e-4647-8b3f-d79137d096eb"). InnerVolumeSpecName "kube-api-access-wmgjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.840280 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 19:03:23 crc kubenswrapper[4733]: E1204 19:03:23.840892 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="177fc63c-9a4e-4647-8b3f-d79137d096eb" containerName="mariadb-client-5-default" Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.840926 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="177fc63c-9a4e-4647-8b3f-d79137d096eb" containerName="mariadb-client-5-default" Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.841273 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="177fc63c-9a4e-4647-8b3f-d79137d096eb" containerName="mariadb-client-5-default" Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.842713 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.855482 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.881213 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmgjr\" (UniqueName: \"kubernetes.io/projected/177fc63c-9a4e-4647-8b3f-d79137d096eb-kube-api-access-wmgjr\") on node \"crc\" DevicePath \"\"" Dec 04 19:03:23 crc kubenswrapper[4733]: I1204 19:03:23.982314 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxh2r\" (UniqueName: \"kubernetes.io/projected/88028d58-8f91-4ac9-a0b3-6b44bfc534e6-kube-api-access-hxh2r\") pod \"mariadb-client-6-default\" (UID: \"88028d58-8f91-4ac9-a0b3-6b44bfc534e6\") " pod="openstack/mariadb-client-6-default" Dec 04 19:03:24 crc kubenswrapper[4733]: I1204 19:03:24.084080 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxh2r\" (UniqueName: \"kubernetes.io/projected/88028d58-8f91-4ac9-a0b3-6b44bfc534e6-kube-api-access-hxh2r\") pod \"mariadb-client-6-default\" (UID: \"88028d58-8f91-4ac9-a0b3-6b44bfc534e6\") " pod="openstack/mariadb-client-6-default" Dec 04 19:03:24 crc kubenswrapper[4733]: I1204 19:03:24.103136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxh2r\" (UniqueName: \"kubernetes.io/projected/88028d58-8f91-4ac9-a0b3-6b44bfc534e6-kube-api-access-hxh2r\") pod \"mariadb-client-6-default\" (UID: \"88028d58-8f91-4ac9-a0b3-6b44bfc534e6\") " pod="openstack/mariadb-client-6-default" Dec 04 19:03:24 crc kubenswrapper[4733]: I1204 19:03:24.161052 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 04 19:03:24 crc kubenswrapper[4733]: I1204 19:03:24.315103 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21ac45a79b5027bd05a91a327f2ec2405d5d1541f48b9aa8c8fe27092a8c0567" Dec 04 19:03:24 crc kubenswrapper[4733]: I1204 19:03:24.315195 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 04 19:03:24 crc kubenswrapper[4733]: I1204 19:03:24.353399 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="177fc63c-9a4e-4647-8b3f-d79137d096eb" path="/var/lib/kubelet/pods/177fc63c-9a4e-4647-8b3f-d79137d096eb/volumes" Dec 04 19:03:24 crc kubenswrapper[4733]: I1204 19:03:24.704641 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 19:03:24 crc kubenswrapper[4733]: W1204 19:03:24.711531 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88028d58_8f91_4ac9_a0b3_6b44bfc534e6.slice/crio-58109269f4bcff91c8f072a68f4edfa45c76ff2d66cfcd34e5801df65432c484 WatchSource:0}: Error finding container 58109269f4bcff91c8f072a68f4edfa45c76ff2d66cfcd34e5801df65432c484: Status 404 returned error can't find the container with id 58109269f4bcff91c8f072a68f4edfa45c76ff2d66cfcd34e5801df65432c484 Dec 04 19:03:25 crc kubenswrapper[4733]: I1204 19:03:25.327139 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"88028d58-8f91-4ac9-a0b3-6b44bfc534e6","Type":"ContainerStarted","Data":"b3a88e8d30c30cd28b19529869ac6cf5cf9fc04d260278f7f7e02cccdc0d2d5a"} Dec 04 19:03:25 crc kubenswrapper[4733]: I1204 19:03:25.327574 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"88028d58-8f91-4ac9-a0b3-6b44bfc534e6","Type":"ContainerStarted","Data":"58109269f4bcff91c8f072a68f4edfa45c76ff2d66cfcd34e5801df65432c484"} Dec 04 19:03:25 crc kubenswrapper[4733]: I1204 19:03:25.350493 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=2.350474723 podStartE2EDuration="2.350474723s" podCreationTimestamp="2025-12-04 19:03:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:03:25.34408489 +0000 UTC m=+5067.299445976" watchObservedRunningTime="2025-12-04 19:03:25.350474723 +0000 UTC m=+5067.305835769" Dec 04 19:03:25 crc kubenswrapper[4733]: I1204 19:03:25.418577 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_88028d58-8f91-4ac9-a0b3-6b44bfc534e6/mariadb-client-6-default/0.log" Dec 04 19:03:26 crc kubenswrapper[4733]: I1204 19:03:26.346340 4733 generic.go:334] "Generic (PLEG): container finished" podID="88028d58-8f91-4ac9-a0b3-6b44bfc534e6" containerID="b3a88e8d30c30cd28b19529869ac6cf5cf9fc04d260278f7f7e02cccdc0d2d5a" exitCode=1 Dec 04 19:03:26 crc kubenswrapper[4733]: I1204 19:03:26.346444 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"88028d58-8f91-4ac9-a0b3-6b44bfc534e6","Type":"ContainerDied","Data":"b3a88e8d30c30cd28b19529869ac6cf5cf9fc04d260278f7f7e02cccdc0d2d5a"} Dec 04 19:03:27 crc kubenswrapper[4733]: I1204 19:03:27.851299 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 04 19:03:27 crc kubenswrapper[4733]: I1204 19:03:27.894174 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 19:03:27 crc kubenswrapper[4733]: I1204 19:03:27.901842 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 19:03:27 crc kubenswrapper[4733]: I1204 19:03:27.951620 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxh2r\" (UniqueName: \"kubernetes.io/projected/88028d58-8f91-4ac9-a0b3-6b44bfc534e6-kube-api-access-hxh2r\") pod \"88028d58-8f91-4ac9-a0b3-6b44bfc534e6\" (UID: \"88028d58-8f91-4ac9-a0b3-6b44bfc534e6\") " Dec 04 19:03:27 crc kubenswrapper[4733]: I1204 19:03:27.960034 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88028d58-8f91-4ac9-a0b3-6b44bfc534e6-kube-api-access-hxh2r" (OuterVolumeSpecName: "kube-api-access-hxh2r") pod "88028d58-8f91-4ac9-a0b3-6b44bfc534e6" (UID: "88028d58-8f91-4ac9-a0b3-6b44bfc534e6"). InnerVolumeSpecName "kube-api-access-hxh2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.053499 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxh2r\" (UniqueName: \"kubernetes.io/projected/88028d58-8f91-4ac9-a0b3-6b44bfc534e6-kube-api-access-hxh2r\") on node \"crc\" DevicePath \"\"" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.076003 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 19:03:28 crc kubenswrapper[4733]: E1204 19:03:28.076769 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88028d58-8f91-4ac9-a0b3-6b44bfc534e6" containerName="mariadb-client-6-default" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.076835 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="88028d58-8f91-4ac9-a0b3-6b44bfc534e6" containerName="mariadb-client-6-default" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.077304 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="88028d58-8f91-4ac9-a0b3-6b44bfc534e6" containerName="mariadb-client-6-default" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.078752 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.083615 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.257521 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j75gc\" (UniqueName: \"kubernetes.io/projected/0a2874d0-b433-418d-bfe8-3f52d1ebcc41-kube-api-access-j75gc\") pod \"mariadb-client-7-default\" (UID: \"0a2874d0-b433-418d-bfe8-3f52d1ebcc41\") " pod="openstack/mariadb-client-7-default" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.351177 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88028d58-8f91-4ac9-a0b3-6b44bfc534e6" path="/var/lib/kubelet/pods/88028d58-8f91-4ac9-a0b3-6b44bfc534e6/volumes" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.358546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j75gc\" (UniqueName: \"kubernetes.io/projected/0a2874d0-b433-418d-bfe8-3f52d1ebcc41-kube-api-access-j75gc\") pod \"mariadb-client-7-default\" (UID: \"0a2874d0-b433-418d-bfe8-3f52d1ebcc41\") " pod="openstack/mariadb-client-7-default" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.370692 4733 scope.go:117] "RemoveContainer" containerID="b3a88e8d30c30cd28b19529869ac6cf5cf9fc04d260278f7f7e02cccdc0d2d5a" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.370772 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.377908 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j75gc\" (UniqueName: \"kubernetes.io/projected/0a2874d0-b433-418d-bfe8-3f52d1ebcc41-kube-api-access-j75gc\") pod \"mariadb-client-7-default\" (UID: \"0a2874d0-b433-418d-bfe8-3f52d1ebcc41\") " pod="openstack/mariadb-client-7-default" Dec 04 19:03:28 crc kubenswrapper[4733]: I1204 19:03:28.402620 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 04 19:03:29 crc kubenswrapper[4733]: I1204 19:03:29.013449 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 19:03:29 crc kubenswrapper[4733]: W1204 19:03:29.016732 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a2874d0_b433_418d_bfe8_3f52d1ebcc41.slice/crio-caea8f3e40a505bdd2779f0eb370553855838758680d836b296df71aeb650e71 WatchSource:0}: Error finding container caea8f3e40a505bdd2779f0eb370553855838758680d836b296df71aeb650e71: Status 404 returned error can't find the container with id caea8f3e40a505bdd2779f0eb370553855838758680d836b296df71aeb650e71 Dec 04 19:03:29 crc kubenswrapper[4733]: I1204 19:03:29.382619 4733 generic.go:334] "Generic (PLEG): container finished" podID="0a2874d0-b433-418d-bfe8-3f52d1ebcc41" containerID="8baccd9d860b414cabb3e3a42c4a3eef8ec9f39ca4a5b88fef7a428dc8b71a1e" exitCode=0 Dec 04 19:03:29 crc kubenswrapper[4733]: I1204 19:03:29.382746 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"0a2874d0-b433-418d-bfe8-3f52d1ebcc41","Type":"ContainerDied","Data":"8baccd9d860b414cabb3e3a42c4a3eef8ec9f39ca4a5b88fef7a428dc8b71a1e"} Dec 04 19:03:29 crc kubenswrapper[4733]: I1204 19:03:29.383106 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"0a2874d0-b433-418d-bfe8-3f52d1ebcc41","Type":"ContainerStarted","Data":"caea8f3e40a505bdd2779f0eb370553855838758680d836b296df71aeb650e71"} Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.767369 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.787878 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_0a2874d0-b433-418d-bfe8-3f52d1ebcc41/mariadb-client-7-default/0.log" Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.820496 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.826286 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.896138 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j75gc\" (UniqueName: \"kubernetes.io/projected/0a2874d0-b433-418d-bfe8-3f52d1ebcc41-kube-api-access-j75gc\") pod \"0a2874d0-b433-418d-bfe8-3f52d1ebcc41\" (UID: \"0a2874d0-b433-418d-bfe8-3f52d1ebcc41\") " Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.901390 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a2874d0-b433-418d-bfe8-3f52d1ebcc41-kube-api-access-j75gc" (OuterVolumeSpecName: "kube-api-access-j75gc") pod "0a2874d0-b433-418d-bfe8-3f52d1ebcc41" (UID: "0a2874d0-b433-418d-bfe8-3f52d1ebcc41"). InnerVolumeSpecName "kube-api-access-j75gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.967857 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Dec 04 19:03:30 crc kubenswrapper[4733]: E1204 19:03:30.968387 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2874d0-b433-418d-bfe8-3f52d1ebcc41" containerName="mariadb-client-7-default" Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.968461 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2874d0-b433-418d-bfe8-3f52d1ebcc41" containerName="mariadb-client-7-default" Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.968721 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2874d0-b433-418d-bfe8-3f52d1ebcc41" containerName="mariadb-client-7-default" Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.969393 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 04 19:03:30 crc kubenswrapper[4733]: I1204 19:03:30.973095 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 04 19:03:31 crc kubenswrapper[4733]: I1204 19:03:31.014478 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j75gc\" (UniqueName: \"kubernetes.io/projected/0a2874d0-b433-418d-bfe8-3f52d1ebcc41-kube-api-access-j75gc\") on node \"crc\" DevicePath \"\"" Dec 04 19:03:31 crc kubenswrapper[4733]: I1204 19:03:31.116093 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-944cs\" (UniqueName: \"kubernetes.io/projected/d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5-kube-api-access-944cs\") pod \"mariadb-client-2\" (UID: \"d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5\") " pod="openstack/mariadb-client-2" Dec 04 19:03:31 crc kubenswrapper[4733]: I1204 19:03:31.217632 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-944cs\" (UniqueName: \"kubernetes.io/projected/d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5-kube-api-access-944cs\") pod \"mariadb-client-2\" (UID: \"d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5\") " pod="openstack/mariadb-client-2" Dec 04 19:03:31 crc kubenswrapper[4733]: I1204 19:03:31.247628 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-944cs\" (UniqueName: \"kubernetes.io/projected/d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5-kube-api-access-944cs\") pod \"mariadb-client-2\" (UID: \"d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5\") " pod="openstack/mariadb-client-2" Dec 04 19:03:31 crc kubenswrapper[4733]: I1204 19:03:31.339659 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 04 19:03:31 crc kubenswrapper[4733]: I1204 19:03:31.415359 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caea8f3e40a505bdd2779f0eb370553855838758680d836b296df71aeb650e71" Dec 04 19:03:31 crc kubenswrapper[4733]: I1204 19:03:31.415437 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 04 19:03:31 crc kubenswrapper[4733]: I1204 19:03:31.895689 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 04 19:03:32 crc kubenswrapper[4733]: I1204 19:03:32.345167 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a2874d0-b433-418d-bfe8-3f52d1ebcc41" path="/var/lib/kubelet/pods/0a2874d0-b433-418d-bfe8-3f52d1ebcc41/volumes" Dec 04 19:03:32 crc kubenswrapper[4733]: I1204 19:03:32.425308 4733 generic.go:334] "Generic (PLEG): container finished" podID="d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5" containerID="0c54dc1af8b92bda535534a9ef92fedd754d5fc773e802e08731ca42289655e9" exitCode=0 Dec 04 19:03:32 crc kubenswrapper[4733]: I1204 19:03:32.425370 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5","Type":"ContainerDied","Data":"0c54dc1af8b92bda535534a9ef92fedd754d5fc773e802e08731ca42289655e9"} Dec 04 19:03:32 crc kubenswrapper[4733]: I1204 19:03:32.425410 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5","Type":"ContainerStarted","Data":"0028e1cbdfe009fea287b4c9f5e7e40d772f5afe10e853fa63fe633c97e1e1e3"} Dec 04 19:03:33 crc kubenswrapper[4733]: I1204 19:03:33.777247 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 04 19:03:33 crc kubenswrapper[4733]: I1204 19:03:33.794973 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5/mariadb-client-2/0.log" Dec 04 19:03:33 crc kubenswrapper[4733]: I1204 19:03:33.824454 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Dec 04 19:03:33 crc kubenswrapper[4733]: I1204 19:03:33.830099 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Dec 04 19:03:33 crc kubenswrapper[4733]: I1204 19:03:33.878916 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-944cs\" (UniqueName: \"kubernetes.io/projected/d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5-kube-api-access-944cs\") pod \"d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5\" (UID: \"d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5\") " Dec 04 19:03:33 crc kubenswrapper[4733]: I1204 19:03:33.885880 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5-kube-api-access-944cs" (OuterVolumeSpecName: "kube-api-access-944cs") pod "d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5" (UID: "d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5"). InnerVolumeSpecName "kube-api-access-944cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:03:33 crc kubenswrapper[4733]: I1204 19:03:33.980267 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-944cs\" (UniqueName: \"kubernetes.io/projected/d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5-kube-api-access-944cs\") on node \"crc\" DevicePath \"\"" Dec 04 19:03:34 crc kubenswrapper[4733]: I1204 19:03:34.357145 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5" path="/var/lib/kubelet/pods/d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5/volumes" Dec 04 19:03:34 crc kubenswrapper[4733]: I1204 19:03:34.450468 4733 scope.go:117] "RemoveContainer" containerID="0c54dc1af8b92bda535534a9ef92fedd754d5fc773e802e08731ca42289655e9" Dec 04 19:03:34 crc kubenswrapper[4733]: I1204 19:03:34.450476 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.362177 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.362919 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.362971 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.363619 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.363692 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" gracePeriod=600 Dec 04 19:03:45 crc kubenswrapper[4733]: E1204 19:03:45.485343 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.559859 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" exitCode=0 Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.559907 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf"} Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.559950 4733 scope.go:117] "RemoveContainer" containerID="133576a4fa20da80bc0ad5a9ec576a91784d7d1b185b32466a081bc4c9e98f7f" Dec 04 19:03:45 crc kubenswrapper[4733]: I1204 19:03:45.560494 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:03:45 crc kubenswrapper[4733]: E1204 19:03:45.560767 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:03:58 crc kubenswrapper[4733]: I1204 19:03:58.351841 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:03:58 crc kubenswrapper[4733]: E1204 19:03:58.352975 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:04:09 crc kubenswrapper[4733]: I1204 19:04:09.336341 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:04:09 crc kubenswrapper[4733]: E1204 19:04:09.337296 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:04:20 crc kubenswrapper[4733]: I1204 19:04:20.335983 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:04:20 crc kubenswrapper[4733]: E1204 19:04:20.336747 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:04:21 crc kubenswrapper[4733]: I1204 19:04:21.049979 4733 scope.go:117] "RemoveContainer" containerID="66241135ddd16a802ad2cb212387ee2705c41a0f0015af7e04ba7cdfd47717cb" Dec 04 19:04:31 crc kubenswrapper[4733]: I1204 19:04:31.336233 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:04:31 crc kubenswrapper[4733]: E1204 19:04:31.337039 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:04:43 crc kubenswrapper[4733]: I1204 19:04:43.335303 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:04:43 crc kubenswrapper[4733]: E1204 19:04:43.336118 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:04:54 crc kubenswrapper[4733]: I1204 19:04:54.336489 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:04:54 crc kubenswrapper[4733]: E1204 19:04:54.337636 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:05:06 crc kubenswrapper[4733]: I1204 19:05:06.335306 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:05:06 crc kubenswrapper[4733]: E1204 19:05:06.336078 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:05:21 crc kubenswrapper[4733]: I1204 19:05:21.335998 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:05:21 crc kubenswrapper[4733]: E1204 19:05:21.337071 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:05:32 crc kubenswrapper[4733]: I1204 19:05:32.336065 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:05:32 crc kubenswrapper[4733]: E1204 19:05:32.336841 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:05:33 crc kubenswrapper[4733]: I1204 19:05:33.738850 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vvsl8"] Dec 04 19:05:33 crc kubenswrapper[4733]: E1204 19:05:33.739553 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5" containerName="mariadb-client-2" Dec 04 19:05:33 crc kubenswrapper[4733]: I1204 19:05:33.739567 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5" containerName="mariadb-client-2" Dec 04 19:05:33 crc kubenswrapper[4733]: I1204 19:05:33.739744 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5f2b76f-cdf4-457e-bb2f-b9ea28881cb5" containerName="mariadb-client-2" Dec 04 19:05:33 crc kubenswrapper[4733]: I1204 19:05:33.740986 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:33 crc kubenswrapper[4733]: I1204 19:05:33.755585 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vvsl8"] Dec 04 19:05:33 crc kubenswrapper[4733]: I1204 19:05:33.920052 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdfc8\" (UniqueName: \"kubernetes.io/projected/35a4ee02-a6d3-4304-9d99-b3f9a371453d-kube-api-access-xdfc8\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:33 crc kubenswrapper[4733]: I1204 19:05:33.920185 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-catalog-content\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:33 crc kubenswrapper[4733]: I1204 19:05:33.920230 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-utilities\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:34 crc kubenswrapper[4733]: I1204 19:05:34.021669 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-utilities\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:34 crc kubenswrapper[4733]: I1204 19:05:34.021825 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdfc8\" (UniqueName: \"kubernetes.io/projected/35a4ee02-a6d3-4304-9d99-b3f9a371453d-kube-api-access-xdfc8\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:34 crc kubenswrapper[4733]: I1204 19:05:34.021883 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-catalog-content\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:34 crc kubenswrapper[4733]: I1204 19:05:34.022281 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-utilities\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:34 crc kubenswrapper[4733]: I1204 19:05:34.022423 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-catalog-content\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:34 crc kubenswrapper[4733]: I1204 19:05:34.055776 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdfc8\" (UniqueName: \"kubernetes.io/projected/35a4ee02-a6d3-4304-9d99-b3f9a371453d-kube-api-access-xdfc8\") pod \"community-operators-vvsl8\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:34 crc kubenswrapper[4733]: I1204 19:05:34.061408 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:34 crc kubenswrapper[4733]: I1204 19:05:34.580508 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vvsl8"] Dec 04 19:05:35 crc kubenswrapper[4733]: I1204 19:05:35.544092 4733 generic.go:334] "Generic (PLEG): container finished" podID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerID="9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409" exitCode=0 Dec 04 19:05:35 crc kubenswrapper[4733]: I1204 19:05:35.544141 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvsl8" event={"ID":"35a4ee02-a6d3-4304-9d99-b3f9a371453d","Type":"ContainerDied","Data":"9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409"} Dec 04 19:05:35 crc kubenswrapper[4733]: I1204 19:05:35.544390 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvsl8" event={"ID":"35a4ee02-a6d3-4304-9d99-b3f9a371453d","Type":"ContainerStarted","Data":"0a19c89ae0107ae28dd553470c7e63c64ce526739f3fe080adebcc67ecb13bef"} Dec 04 19:05:36 crc kubenswrapper[4733]: I1204 19:05:36.554445 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvsl8" event={"ID":"35a4ee02-a6d3-4304-9d99-b3f9a371453d","Type":"ContainerStarted","Data":"939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da"} Dec 04 19:05:37 crc kubenswrapper[4733]: I1204 19:05:37.565812 4733 generic.go:334] "Generic (PLEG): container finished" podID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerID="939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da" exitCode=0 Dec 04 19:05:37 crc kubenswrapper[4733]: I1204 19:05:37.565909 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvsl8" event={"ID":"35a4ee02-a6d3-4304-9d99-b3f9a371453d","Type":"ContainerDied","Data":"939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da"} Dec 04 19:05:38 crc kubenswrapper[4733]: I1204 19:05:38.578134 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvsl8" event={"ID":"35a4ee02-a6d3-4304-9d99-b3f9a371453d","Type":"ContainerStarted","Data":"1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186"} Dec 04 19:05:38 crc kubenswrapper[4733]: I1204 19:05:38.608839 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vvsl8" podStartSLOduration=3.192314595 podStartE2EDuration="5.608819762s" podCreationTimestamp="2025-12-04 19:05:33 +0000 UTC" firstStartedPulling="2025-12-04 19:05:35.54637179 +0000 UTC m=+5197.501732836" lastFinishedPulling="2025-12-04 19:05:37.962876947 +0000 UTC m=+5199.918238003" observedRunningTime="2025-12-04 19:05:38.60173104 +0000 UTC m=+5200.557092126" watchObservedRunningTime="2025-12-04 19:05:38.608819762 +0000 UTC m=+5200.564180818" Dec 04 19:05:44 crc kubenswrapper[4733]: I1204 19:05:44.061844 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:44 crc kubenswrapper[4733]: I1204 19:05:44.063010 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:44 crc kubenswrapper[4733]: I1204 19:05:44.105035 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:44 crc kubenswrapper[4733]: I1204 19:05:44.702455 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:44 crc kubenswrapper[4733]: I1204 19:05:44.761177 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vvsl8"] Dec 04 19:05:45 crc kubenswrapper[4733]: I1204 19:05:45.335328 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:05:45 crc kubenswrapper[4733]: E1204 19:05:45.335541 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:05:46 crc kubenswrapper[4733]: I1204 19:05:46.660042 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vvsl8" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerName="registry-server" containerID="cri-o://1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186" gracePeriod=2 Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.189958 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.331391 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdfc8\" (UniqueName: \"kubernetes.io/projected/35a4ee02-a6d3-4304-9d99-b3f9a371453d-kube-api-access-xdfc8\") pod \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.331563 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-catalog-content\") pod \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.331633 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-utilities\") pod \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\" (UID: \"35a4ee02-a6d3-4304-9d99-b3f9a371453d\") " Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.332820 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-utilities" (OuterVolumeSpecName: "utilities") pod "35a4ee02-a6d3-4304-9d99-b3f9a371453d" (UID: "35a4ee02-a6d3-4304-9d99-b3f9a371453d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.359070 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35a4ee02-a6d3-4304-9d99-b3f9a371453d-kube-api-access-xdfc8" (OuterVolumeSpecName: "kube-api-access-xdfc8") pod "35a4ee02-a6d3-4304-9d99-b3f9a371453d" (UID: "35a4ee02-a6d3-4304-9d99-b3f9a371453d"). InnerVolumeSpecName "kube-api-access-xdfc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.416776 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35a4ee02-a6d3-4304-9d99-b3f9a371453d" (UID: "35a4ee02-a6d3-4304-9d99-b3f9a371453d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.433776 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdfc8\" (UniqueName: \"kubernetes.io/projected/35a4ee02-a6d3-4304-9d99-b3f9a371453d-kube-api-access-xdfc8\") on node \"crc\" DevicePath \"\"" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.433831 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.433843 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4ee02-a6d3-4304-9d99-b3f9a371453d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.668224 4733 generic.go:334] "Generic (PLEG): container finished" podID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerID="1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186" exitCode=0 Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.668270 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvsl8" event={"ID":"35a4ee02-a6d3-4304-9d99-b3f9a371453d","Type":"ContainerDied","Data":"1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186"} Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.668316 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvsl8" event={"ID":"35a4ee02-a6d3-4304-9d99-b3f9a371453d","Type":"ContainerDied","Data":"0a19c89ae0107ae28dd553470c7e63c64ce526739f3fe080adebcc67ecb13bef"} Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.668336 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvsl8" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.668362 4733 scope.go:117] "RemoveContainer" containerID="1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.688735 4733 scope.go:117] "RemoveContainer" containerID="939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.701111 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vvsl8"] Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.707664 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vvsl8"] Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.731071 4733 scope.go:117] "RemoveContainer" containerID="9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.749098 4733 scope.go:117] "RemoveContainer" containerID="1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186" Dec 04 19:05:47 crc kubenswrapper[4733]: E1204 19:05:47.749592 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186\": container with ID starting with 1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186 not found: ID does not exist" containerID="1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.749638 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186"} err="failed to get container status \"1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186\": rpc error: code = NotFound desc = could not find container \"1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186\": container with ID starting with 1d6014a07d452240cd9d87756e5ff9916f7876fe8fd929509a73b6a112e85186 not found: ID does not exist" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.749675 4733 scope.go:117] "RemoveContainer" containerID="939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da" Dec 04 19:05:47 crc kubenswrapper[4733]: E1204 19:05:47.750209 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da\": container with ID starting with 939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da not found: ID does not exist" containerID="939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.750243 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da"} err="failed to get container status \"939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da\": rpc error: code = NotFound desc = could not find container \"939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da\": container with ID starting with 939225bdf14b982093559f692e025ca33705d40d61ccc2ea5597a3ee1514a8da not found: ID does not exist" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.750265 4733 scope.go:117] "RemoveContainer" containerID="9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409" Dec 04 19:05:47 crc kubenswrapper[4733]: E1204 19:05:47.750726 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409\": container with ID starting with 9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409 not found: ID does not exist" containerID="9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409" Dec 04 19:05:47 crc kubenswrapper[4733]: I1204 19:05:47.750756 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409"} err="failed to get container status \"9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409\": rpc error: code = NotFound desc = could not find container \"9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409\": container with ID starting with 9db3c6cc5e779bf7009c7f245f5f0a9753fe0ab589ed9ecaa617e6812a6ae409 not found: ID does not exist" Dec 04 19:05:48 crc kubenswrapper[4733]: I1204 19:05:48.346279 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" path="/var/lib/kubelet/pods/35a4ee02-a6d3-4304-9d99-b3f9a371453d/volumes" Dec 04 19:05:58 crc kubenswrapper[4733]: I1204 19:05:58.340738 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:05:58 crc kubenswrapper[4733]: E1204 19:05:58.341648 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:06:09 crc kubenswrapper[4733]: I1204 19:06:09.336171 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:06:09 crc kubenswrapper[4733]: E1204 19:06:09.337309 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:06:21 crc kubenswrapper[4733]: I1204 19:06:21.335339 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:06:21 crc kubenswrapper[4733]: E1204 19:06:21.336295 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:06:36 crc kubenswrapper[4733]: I1204 19:06:36.335669 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:06:36 crc kubenswrapper[4733]: E1204 19:06:36.336668 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:06:48 crc kubenswrapper[4733]: I1204 19:06:48.340640 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:06:48 crc kubenswrapper[4733]: E1204 19:06:48.341884 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:06:59 crc kubenswrapper[4733]: I1204 19:06:59.336565 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:06:59 crc kubenswrapper[4733]: E1204 19:06:59.337493 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:07:11 crc kubenswrapper[4733]: I1204 19:07:11.335172 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:07:11 crc kubenswrapper[4733]: E1204 19:07:11.336011 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:07:23 crc kubenswrapper[4733]: I1204 19:07:23.336846 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:07:23 crc kubenswrapper[4733]: E1204 19:07:23.337953 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:07:38 crc kubenswrapper[4733]: I1204 19:07:38.344310 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:07:38 crc kubenswrapper[4733]: E1204 19:07:38.345515 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:07:49 crc kubenswrapper[4733]: I1204 19:07:49.336212 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:07:49 crc kubenswrapper[4733]: E1204 19:07:49.337301 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:08:04 crc kubenswrapper[4733]: I1204 19:08:04.335667 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:08:04 crc kubenswrapper[4733]: E1204 19:08:04.336585 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.568473 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 19:08:10 crc kubenswrapper[4733]: E1204 19:08:10.569708 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerName="registry-server" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.569755 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerName="registry-server" Dec 04 19:08:10 crc kubenswrapper[4733]: E1204 19:08:10.569788 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerName="extract-content" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.569811 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerName="extract-content" Dec 04 19:08:10 crc kubenswrapper[4733]: E1204 19:08:10.569835 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerName="extract-utilities" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.569841 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerName="extract-utilities" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.570012 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="35a4ee02-a6d3-4304-9d99-b3f9a371453d" containerName="registry-server" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.570644 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.575654 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-hvvvj" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.581193 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.766368 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\") pod \"mariadb-copy-data\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") " pod="openstack/mariadb-copy-data" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.766476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkckc\" (UniqueName: \"kubernetes.io/projected/54a0573d-d1fc-4355-a4e6-96a9db7a1a8f-kube-api-access-vkckc\") pod \"mariadb-copy-data\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") " pod="openstack/mariadb-copy-data" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.867466 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\") pod \"mariadb-copy-data\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") " pod="openstack/mariadb-copy-data" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.867517 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkckc\" (UniqueName: \"kubernetes.io/projected/54a0573d-d1fc-4355-a4e6-96a9db7a1a8f-kube-api-access-vkckc\") pod \"mariadb-copy-data\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") " pod="openstack/mariadb-copy-data" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.872732 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.872775 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\") pod \"mariadb-copy-data\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3cb72c707bb7afd1b36b978a9dfab72724c0c099cef160eb22f5f828755b3fa3/globalmount\"" pod="openstack/mariadb-copy-data" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.906656 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkckc\" (UniqueName: \"kubernetes.io/projected/54a0573d-d1fc-4355-a4e6-96a9db7a1a8f-kube-api-access-vkckc\") pod \"mariadb-copy-data\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") " pod="openstack/mariadb-copy-data" Dec 04 19:08:10 crc kubenswrapper[4733]: I1204 19:08:10.916143 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\") pod \"mariadb-copy-data\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") " pod="openstack/mariadb-copy-data" Dec 04 19:08:11 crc kubenswrapper[4733]: I1204 19:08:11.194626 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 04 19:08:11 crc kubenswrapper[4733]: I1204 19:08:11.779225 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 19:08:11 crc kubenswrapper[4733]: I1204 19:08:11.957332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f","Type":"ContainerStarted","Data":"af26d3bd6fcd8d1d31a7631394b5571a3ff799e96742e0ce6e5afa2164a0f752"} Dec 04 19:08:12 crc kubenswrapper[4733]: I1204 19:08:12.970768 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f","Type":"ContainerStarted","Data":"b108cc8a169835be119998e8931dce65f70fc8422eca3e347c388d28f51b1613"} Dec 04 19:08:12 crc kubenswrapper[4733]: I1204 19:08:12.993891 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.993869548 podStartE2EDuration="3.993869548s" podCreationTimestamp="2025-12-04 19:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:08:12.987416174 +0000 UTC m=+5354.942777270" watchObservedRunningTime="2025-12-04 19:08:12.993869548 +0000 UTC m=+5354.949230614" Dec 04 19:08:15 crc kubenswrapper[4733]: I1204 19:08:15.337156 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:08:15 crc kubenswrapper[4733]: E1204 19:08:15.338095 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:08:15 crc kubenswrapper[4733]: I1204 19:08:15.935502 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:15 crc kubenswrapper[4733]: I1204 19:08:15.936395 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 19:08:15 crc kubenswrapper[4733]: I1204 19:08:15.951290 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:16 crc kubenswrapper[4733]: I1204 19:08:16.050613 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj4nv\" (UniqueName: \"kubernetes.io/projected/c78f0960-1eec-493f-910e-14683cc1c9a8-kube-api-access-bj4nv\") pod \"mariadb-client\" (UID: \"c78f0960-1eec-493f-910e-14683cc1c9a8\") " pod="openstack/mariadb-client" Dec 04 19:08:16 crc kubenswrapper[4733]: I1204 19:08:16.152029 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj4nv\" (UniqueName: \"kubernetes.io/projected/c78f0960-1eec-493f-910e-14683cc1c9a8-kube-api-access-bj4nv\") pod \"mariadb-client\" (UID: \"c78f0960-1eec-493f-910e-14683cc1c9a8\") " pod="openstack/mariadb-client" Dec 04 19:08:16 crc kubenswrapper[4733]: I1204 19:08:16.173710 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj4nv\" (UniqueName: \"kubernetes.io/projected/c78f0960-1eec-493f-910e-14683cc1c9a8-kube-api-access-bj4nv\") pod \"mariadb-client\" (UID: \"c78f0960-1eec-493f-910e-14683cc1c9a8\") " pod="openstack/mariadb-client" Dec 04 19:08:16 crc kubenswrapper[4733]: I1204 19:08:16.256868 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 19:08:16 crc kubenswrapper[4733]: I1204 19:08:16.725360 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:16 crc kubenswrapper[4733]: W1204 19:08:16.734101 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc78f0960_1eec_493f_910e_14683cc1c9a8.slice/crio-534bfeac5f038ac99e9c91f7cef0bba3b67aec698ceecda44c3dfafe7e3cfd74 WatchSource:0}: Error finding container 534bfeac5f038ac99e9c91f7cef0bba3b67aec698ceecda44c3dfafe7e3cfd74: Status 404 returned error can't find the container with id 534bfeac5f038ac99e9c91f7cef0bba3b67aec698ceecda44c3dfafe7e3cfd74 Dec 04 19:08:17 crc kubenswrapper[4733]: I1204 19:08:17.005513 4733 generic.go:334] "Generic (PLEG): container finished" podID="c78f0960-1eec-493f-910e-14683cc1c9a8" containerID="a780193699f74ba0880403e8a212a23ba5e0a63948555b48c044896dc9b5e7a7" exitCode=0 Dec 04 19:08:17 crc kubenswrapper[4733]: I1204 19:08:17.005571 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c78f0960-1eec-493f-910e-14683cc1c9a8","Type":"ContainerDied","Data":"a780193699f74ba0880403e8a212a23ba5e0a63948555b48c044896dc9b5e7a7"} Dec 04 19:08:17 crc kubenswrapper[4733]: I1204 19:08:17.005790 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c78f0960-1eec-493f-910e-14683cc1c9a8","Type":"ContainerStarted","Data":"534bfeac5f038ac99e9c91f7cef0bba3b67aec698ceecda44c3dfafe7e3cfd74"} Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.421705 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.450909 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_c78f0960-1eec-493f-910e-14683cc1c9a8/mariadb-client/0.log" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.489927 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.498106 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.592629 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj4nv\" (UniqueName: \"kubernetes.io/projected/c78f0960-1eec-493f-910e-14683cc1c9a8-kube-api-access-bj4nv\") pod \"c78f0960-1eec-493f-910e-14683cc1c9a8\" (UID: \"c78f0960-1eec-493f-910e-14683cc1c9a8\") " Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.603164 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c78f0960-1eec-493f-910e-14683cc1c9a8-kube-api-access-bj4nv" (OuterVolumeSpecName: "kube-api-access-bj4nv") pod "c78f0960-1eec-493f-910e-14683cc1c9a8" (UID: "c78f0960-1eec-493f-910e-14683cc1c9a8"). InnerVolumeSpecName "kube-api-access-bj4nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.687623 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:18 crc kubenswrapper[4733]: E1204 19:08:18.688006 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c78f0960-1eec-493f-910e-14683cc1c9a8" containerName="mariadb-client" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.688032 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78f0960-1eec-493f-910e-14683cc1c9a8" containerName="mariadb-client" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.688236 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c78f0960-1eec-493f-910e-14683cc1c9a8" containerName="mariadb-client" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.688855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.696122 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj4nv\" (UniqueName: \"kubernetes.io/projected/c78f0960-1eec-493f-910e-14683cc1c9a8-kube-api-access-bj4nv\") on node \"crc\" DevicePath \"\"" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.713646 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.798192 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtltl\" (UniqueName: \"kubernetes.io/projected/b3c9f0e2-151e-4ce1-a096-67df2162b376-kube-api-access-dtltl\") pod \"mariadb-client\" (UID: \"b3c9f0e2-151e-4ce1-a096-67df2162b376\") " pod="openstack/mariadb-client" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.900033 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtltl\" (UniqueName: \"kubernetes.io/projected/b3c9f0e2-151e-4ce1-a096-67df2162b376-kube-api-access-dtltl\") pod \"mariadb-client\" (UID: \"b3c9f0e2-151e-4ce1-a096-67df2162b376\") " pod="openstack/mariadb-client" Dec 04 19:08:18 crc kubenswrapper[4733]: I1204 19:08:18.921334 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtltl\" (UniqueName: \"kubernetes.io/projected/b3c9f0e2-151e-4ce1-a096-67df2162b376-kube-api-access-dtltl\") pod \"mariadb-client\" (UID: \"b3c9f0e2-151e-4ce1-a096-67df2162b376\") " pod="openstack/mariadb-client" Dec 04 19:08:19 crc kubenswrapper[4733]: I1204 19:08:19.035982 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="534bfeac5f038ac99e9c91f7cef0bba3b67aec698ceecda44c3dfafe7e3cfd74" Dec 04 19:08:19 crc kubenswrapper[4733]: I1204 19:08:19.036076 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 19:08:19 crc kubenswrapper[4733]: I1204 19:08:19.038413 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 19:08:19 crc kubenswrapper[4733]: I1204 19:08:19.111592 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="c78f0960-1eec-493f-910e-14683cc1c9a8" podUID="b3c9f0e2-151e-4ce1-a096-67df2162b376" Dec 04 19:08:19 crc kubenswrapper[4733]: I1204 19:08:19.531131 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:20 crc kubenswrapper[4733]: I1204 19:08:20.046405 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"b3c9f0e2-151e-4ce1-a096-67df2162b376","Type":"ContainerStarted","Data":"8cbae04a246503e32faae2b34749a16e6406be5b917158daa02ae4ef98480f4a"} Dec 04 19:08:20 crc kubenswrapper[4733]: I1204 19:08:20.348244 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c78f0960-1eec-493f-910e-14683cc1c9a8" path="/var/lib/kubelet/pods/c78f0960-1eec-493f-910e-14683cc1c9a8/volumes" Dec 04 19:08:21 crc kubenswrapper[4733]: I1204 19:08:21.061068 4733 generic.go:334] "Generic (PLEG): container finished" podID="b3c9f0e2-151e-4ce1-a096-67df2162b376" containerID="6fe2c2d7f492eeeceb56fe6f49ed43167571c3108bb54338838590bf5c59c7bb" exitCode=0 Dec 04 19:08:21 crc kubenswrapper[4733]: I1204 19:08:21.061117 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"b3c9f0e2-151e-4ce1-a096-67df2162b376","Type":"ContainerDied","Data":"6fe2c2d7f492eeeceb56fe6f49ed43167571c3108bb54338838590bf5c59c7bb"} Dec 04 19:08:22 crc kubenswrapper[4733]: I1204 19:08:22.406585 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 19:08:22 crc kubenswrapper[4733]: I1204 19:08:22.429765 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_b3c9f0e2-151e-4ce1-a096-67df2162b376/mariadb-client/0.log" Dec 04 19:08:22 crc kubenswrapper[4733]: I1204 19:08:22.453519 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtltl\" (UniqueName: \"kubernetes.io/projected/b3c9f0e2-151e-4ce1-a096-67df2162b376-kube-api-access-dtltl\") pod \"b3c9f0e2-151e-4ce1-a096-67df2162b376\" (UID: \"b3c9f0e2-151e-4ce1-a096-67df2162b376\") " Dec 04 19:08:22 crc kubenswrapper[4733]: I1204 19:08:22.456959 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:22 crc kubenswrapper[4733]: I1204 19:08:22.459783 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c9f0e2-151e-4ce1-a096-67df2162b376-kube-api-access-dtltl" (OuterVolumeSpecName: "kube-api-access-dtltl") pod "b3c9f0e2-151e-4ce1-a096-67df2162b376" (UID: "b3c9f0e2-151e-4ce1-a096-67df2162b376"). InnerVolumeSpecName "kube-api-access-dtltl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:08:22 crc kubenswrapper[4733]: I1204 19:08:22.464830 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 04 19:08:22 crc kubenswrapper[4733]: I1204 19:08:22.555993 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtltl\" (UniqueName: \"kubernetes.io/projected/b3c9f0e2-151e-4ce1-a096-67df2162b376-kube-api-access-dtltl\") on node \"crc\" DevicePath \"\"" Dec 04 19:08:23 crc kubenswrapper[4733]: I1204 19:08:23.081869 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cbae04a246503e32faae2b34749a16e6406be5b917158daa02ae4ef98480f4a" Dec 04 19:08:23 crc kubenswrapper[4733]: I1204 19:08:23.082010 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 19:08:24 crc kubenswrapper[4733]: I1204 19:08:24.369559 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3c9f0e2-151e-4ce1-a096-67df2162b376" path="/var/lib/kubelet/pods/b3c9f0e2-151e-4ce1-a096-67df2162b376/volumes" Dec 04 19:08:28 crc kubenswrapper[4733]: I1204 19:08:28.339298 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:08:28 crc kubenswrapper[4733]: E1204 19:08:28.339929 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.576276 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gsrhd"] Dec 04 19:08:33 crc kubenswrapper[4733]: E1204 19:08:33.577734 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c9f0e2-151e-4ce1-a096-67df2162b376" containerName="mariadb-client" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.577770 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c9f0e2-151e-4ce1-a096-67df2162b376" containerName="mariadb-client" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.578615 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c9f0e2-151e-4ce1-a096-67df2162b376" containerName="mariadb-client" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.581299 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.591068 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gsrhd"] Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.624363 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vpwp\" (UniqueName: \"kubernetes.io/projected/faefff68-5d9b-4191-929e-937b7b2f057e-kube-api-access-7vpwp\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.624462 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-catalog-content\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.624500 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-utilities\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.731110 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vpwp\" (UniqueName: \"kubernetes.io/projected/faefff68-5d9b-4191-929e-937b7b2f057e-kube-api-access-7vpwp\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.731214 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-catalog-content\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.731257 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-utilities\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.731697 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-catalog-content\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.731910 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-utilities\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.755119 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vpwp\" (UniqueName: \"kubernetes.io/projected/faefff68-5d9b-4191-929e-937b7b2f057e-kube-api-access-7vpwp\") pod \"redhat-operators-gsrhd\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:33 crc kubenswrapper[4733]: I1204 19:08:33.916158 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:34 crc kubenswrapper[4733]: I1204 19:08:34.386502 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gsrhd"] Dec 04 19:08:35 crc kubenswrapper[4733]: I1204 19:08:35.190929 4733 generic.go:334] "Generic (PLEG): container finished" podID="faefff68-5d9b-4191-929e-937b7b2f057e" containerID="c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b" exitCode=0 Dec 04 19:08:35 crc kubenswrapper[4733]: I1204 19:08:35.191249 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsrhd" event={"ID":"faefff68-5d9b-4191-929e-937b7b2f057e","Type":"ContainerDied","Data":"c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b"} Dec 04 19:08:35 crc kubenswrapper[4733]: I1204 19:08:35.192711 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsrhd" event={"ID":"faefff68-5d9b-4191-929e-937b7b2f057e","Type":"ContainerStarted","Data":"2544822a662e1b738db86dbdcf7fd766b47c1a0f38fa7d69feac59913a4d3f47"} Dec 04 19:08:35 crc kubenswrapper[4733]: I1204 19:08:35.195172 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:08:41 crc kubenswrapper[4733]: I1204 19:08:41.337128 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:08:41 crc kubenswrapper[4733]: E1204 19:08:41.338264 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:08:42 crc kubenswrapper[4733]: I1204 19:08:42.292218 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsrhd" event={"ID":"faefff68-5d9b-4191-929e-937b7b2f057e","Type":"ContainerStarted","Data":"021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3"} Dec 04 19:08:43 crc kubenswrapper[4733]: I1204 19:08:43.300127 4733 generic.go:334] "Generic (PLEG): container finished" podID="faefff68-5d9b-4191-929e-937b7b2f057e" containerID="021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3" exitCode=0 Dec 04 19:08:43 crc kubenswrapper[4733]: I1204 19:08:43.300177 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsrhd" event={"ID":"faefff68-5d9b-4191-929e-937b7b2f057e","Type":"ContainerDied","Data":"021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3"} Dec 04 19:08:44 crc kubenswrapper[4733]: I1204 19:08:44.314221 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsrhd" event={"ID":"faefff68-5d9b-4191-929e-937b7b2f057e","Type":"ContainerStarted","Data":"f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea"} Dec 04 19:08:44 crc kubenswrapper[4733]: I1204 19:08:44.337062 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gsrhd" podStartSLOduration=2.618249498 podStartE2EDuration="11.337033245s" podCreationTimestamp="2025-12-04 19:08:33 +0000 UTC" firstStartedPulling="2025-12-04 19:08:35.194625215 +0000 UTC m=+5377.149986301" lastFinishedPulling="2025-12-04 19:08:43.913409002 +0000 UTC m=+5385.868770048" observedRunningTime="2025-12-04 19:08:44.334904048 +0000 UTC m=+5386.290265124" watchObservedRunningTime="2025-12-04 19:08:44.337033245 +0000 UTC m=+5386.292394331" Dec 04 19:08:53 crc kubenswrapper[4733]: I1204 19:08:53.917047 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:53 crc kubenswrapper[4733]: I1204 19:08:53.917709 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:53 crc kubenswrapper[4733]: I1204 19:08:53.977996 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:54 crc kubenswrapper[4733]: I1204 19:08:54.335000 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:08:54 crc kubenswrapper[4733]: I1204 19:08:54.445639 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:08:54 crc kubenswrapper[4733]: I1204 19:08:54.522786 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gsrhd"] Dec 04 19:08:54 crc kubenswrapper[4733]: I1204 19:08:54.570176 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-npgrs"] Dec 04 19:08:54 crc kubenswrapper[4733]: I1204 19:08:54.570458 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-npgrs" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="registry-server" containerID="cri-o://9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c" gracePeriod=2 Dec 04 19:08:55 crc kubenswrapper[4733]: I1204 19:08:55.401385 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"ef4c43597b7d9b6a5f9198424285408b276a51a3896ec710afa65d9b6bbf8b2b"} Dec 04 19:08:56 crc kubenswrapper[4733]: E1204 19:08:56.260068 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c is running failed: container process not found" containerID="9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 19:08:56 crc kubenswrapper[4733]: E1204 19:08:56.261928 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c is running failed: container process not found" containerID="9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 19:08:56 crc kubenswrapper[4733]: E1204 19:08:56.262247 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c is running failed: container process not found" containerID="9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 19:08:56 crc kubenswrapper[4733]: E1204 19:08:56.262292 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-npgrs" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="registry-server" Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.417644 4733 generic.go:334] "Generic (PLEG): container finished" podID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerID="9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c" exitCode=0 Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.417722 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npgrs" event={"ID":"f455a245-3a28-48c0-b83a-5b5b56c897f6","Type":"ContainerDied","Data":"9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c"} Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.733268 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.814481 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46w65\" (UniqueName: \"kubernetes.io/projected/f455a245-3a28-48c0-b83a-5b5b56c897f6-kube-api-access-46w65\") pod \"f455a245-3a28-48c0-b83a-5b5b56c897f6\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.814885 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-utilities\") pod \"f455a245-3a28-48c0-b83a-5b5b56c897f6\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.815008 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-catalog-content\") pod \"f455a245-3a28-48c0-b83a-5b5b56c897f6\" (UID: \"f455a245-3a28-48c0-b83a-5b5b56c897f6\") " Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.815374 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-utilities" (OuterVolumeSpecName: "utilities") pod "f455a245-3a28-48c0-b83a-5b5b56c897f6" (UID: "f455a245-3a28-48c0-b83a-5b5b56c897f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.826065 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f455a245-3a28-48c0-b83a-5b5b56c897f6-kube-api-access-46w65" (OuterVolumeSpecName: "kube-api-access-46w65") pod "f455a245-3a28-48c0-b83a-5b5b56c897f6" (UID: "f455a245-3a28-48c0-b83a-5b5b56c897f6"). InnerVolumeSpecName "kube-api-access-46w65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.902594 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f455a245-3a28-48c0-b83a-5b5b56c897f6" (UID: "f455a245-3a28-48c0-b83a-5b5b56c897f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.916933 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.916967 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46w65\" (UniqueName: \"kubernetes.io/projected/f455a245-3a28-48c0-b83a-5b5b56c897f6-kube-api-access-46w65\") on node \"crc\" DevicePath \"\"" Dec 04 19:08:57 crc kubenswrapper[4733]: I1204 19:08:57.916979 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f455a245-3a28-48c0-b83a-5b5b56c897f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:08:58 crc kubenswrapper[4733]: I1204 19:08:58.427267 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npgrs" event={"ID":"f455a245-3a28-48c0-b83a-5b5b56c897f6","Type":"ContainerDied","Data":"783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4"} Dec 04 19:08:58 crc kubenswrapper[4733]: I1204 19:08:58.427328 4733 scope.go:117] "RemoveContainer" containerID="9ad26a4d311e63cba71a07642c2dbdad66e481cbcd857ae9721f8eec99f6d52c" Dec 04 19:08:58 crc kubenswrapper[4733]: I1204 19:08:58.427491 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npgrs" Dec 04 19:08:58 crc kubenswrapper[4733]: I1204 19:08:58.457142 4733 scope.go:117] "RemoveContainer" containerID="2e7f897c5b3b9347af79d9cea190c633bca6e27299b06ceab8259937b570a9f4" Dec 04 19:08:58 crc kubenswrapper[4733]: I1204 19:08:58.461357 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-npgrs"] Dec 04 19:08:58 crc kubenswrapper[4733]: I1204 19:08:58.472041 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-npgrs"] Dec 04 19:08:58 crc kubenswrapper[4733]: I1204 19:08:58.501239 4733 scope.go:117] "RemoveContainer" containerID="bd83ef32350485919bf0a69613a5a0017daaf6be41e8e9d03557fece45e9861f" Dec 04 19:08:59 crc kubenswrapper[4733]: E1204 19:08:59.842879 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice/crio-783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice\": RecentStats: unable to find data in memory cache]" Dec 04 19:09:00 crc kubenswrapper[4733]: I1204 19:09:00.352999 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" path="/var/lib/kubelet/pods/f455a245-3a28-48c0-b83a-5b5b56c897f6/volumes" Dec 04 19:09:10 crc kubenswrapper[4733]: E1204 19:09:10.067294 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice/crio-783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4\": RecentStats: unable to find data in memory cache]" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.487211 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b7wmd"] Dec 04 19:09:20 crc kubenswrapper[4733]: E1204 19:09:20.488023 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="extract-content" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.488038 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="extract-content" Dec 04 19:09:20 crc kubenswrapper[4733]: E1204 19:09:20.488061 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="extract-utilities" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.488068 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="extract-utilities" Dec 04 19:09:20 crc kubenswrapper[4733]: E1204 19:09:20.488097 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="registry-server" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.488102 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="registry-server" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.488239 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f455a245-3a28-48c0-b83a-5b5b56c897f6" containerName="registry-server" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.491657 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.501529 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b7wmd"] Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.676295 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-catalog-content\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.676406 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9vln\" (UniqueName: \"kubernetes.io/projected/d4955edf-ded2-45c8-9207-c319e9356683-kube-api-access-q9vln\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.676759 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-utilities\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: E1204 19:09:20.685310 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice/crio-783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice\": RecentStats: unable to find data in memory cache]" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.778130 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-utilities\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.778474 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-catalog-content\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.778509 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9vln\" (UniqueName: \"kubernetes.io/projected/d4955edf-ded2-45c8-9207-c319e9356683-kube-api-access-q9vln\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.778744 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-utilities\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.779012 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-catalog-content\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.822495 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9vln\" (UniqueName: \"kubernetes.io/projected/d4955edf-ded2-45c8-9207-c319e9356683-kube-api-access-q9vln\") pod \"certified-operators-b7wmd\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:20 crc kubenswrapper[4733]: I1204 19:09:20.838278 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.223172 4733 scope.go:117] "RemoveContainer" containerID="135aa5524cf9b40263a7992c7090b76cb026c548da8822eb31061e3ed969925c" Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.249489 4733 scope.go:117] "RemoveContainer" containerID="8562146b562bce7b7de00cc9488cc6cff9cbaf5997b4d0ac78a6f884b0997ea2" Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.285672 4733 scope.go:117] "RemoveContainer" containerID="e8a595f3dde8366c3cad95b7ae6cd3edc07217fb3dbcf1d87a183f1f0b58fdcc" Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.305473 4733 scope.go:117] "RemoveContainer" containerID="3c41b451e4b490edb182188de742cb0019132e410db926c309e2c1c94a633552" Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.314238 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b7wmd"] Dec 04 19:09:21 crc kubenswrapper[4733]: W1204 19:09:21.320263 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4955edf_ded2_45c8_9207_c319e9356683.slice/crio-559d5e9a01e2d33bccc497d92f56d635ac5c615d2f7cfbe7e01b82346d8038f7 WatchSource:0}: Error finding container 559d5e9a01e2d33bccc497d92f56d635ac5c615d2f7cfbe7e01b82346d8038f7: Status 404 returned error can't find the container with id 559d5e9a01e2d33bccc497d92f56d635ac5c615d2f7cfbe7e01b82346d8038f7 Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.341904 4733 scope.go:117] "RemoveContainer" containerID="4e7afb3400424199209584a0266d44aa3af1590be06717fffc2b54a8d8dd38a1" Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.364989 4733 scope.go:117] "RemoveContainer" containerID="430eecf1f6573ad992ee7c3b17e2ab6af39b0e71ff55d2c3a55cbc16d483fd30" Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.381399 4733 scope.go:117] "RemoveContainer" containerID="8e9944637059f44dd22dd9c6dddac0be3c602e2776f7be8c448be57e4255534e" Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.688733 4733 generic.go:334] "Generic (PLEG): container finished" podID="d4955edf-ded2-45c8-9207-c319e9356683" containerID="81f3e36d6e0ad74c69e6a565d010513a0ab5f37fc445e0fddf5c6086321d4918" exitCode=0 Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.688783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b7wmd" event={"ID":"d4955edf-ded2-45c8-9207-c319e9356683","Type":"ContainerDied","Data":"81f3e36d6e0ad74c69e6a565d010513a0ab5f37fc445e0fddf5c6086321d4918"} Dec 04 19:09:21 crc kubenswrapper[4733]: I1204 19:09:21.689114 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b7wmd" event={"ID":"d4955edf-ded2-45c8-9207-c319e9356683","Type":"ContainerStarted","Data":"559d5e9a01e2d33bccc497d92f56d635ac5c615d2f7cfbe7e01b82346d8038f7"} Dec 04 19:09:22 crc kubenswrapper[4733]: I1204 19:09:22.698542 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b7wmd" event={"ID":"d4955edf-ded2-45c8-9207-c319e9356683","Type":"ContainerStarted","Data":"ed31019c4114fdafcdab7e1831696b2e6b08dc2c00e913c1eee148e5a17ffd6d"} Dec 04 19:09:23 crc kubenswrapper[4733]: I1204 19:09:23.707966 4733 generic.go:334] "Generic (PLEG): container finished" podID="d4955edf-ded2-45c8-9207-c319e9356683" containerID="ed31019c4114fdafcdab7e1831696b2e6b08dc2c00e913c1eee148e5a17ffd6d" exitCode=0 Dec 04 19:09:23 crc kubenswrapper[4733]: I1204 19:09:23.708035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b7wmd" event={"ID":"d4955edf-ded2-45c8-9207-c319e9356683","Type":"ContainerDied","Data":"ed31019c4114fdafcdab7e1831696b2e6b08dc2c00e913c1eee148e5a17ffd6d"} Dec 04 19:09:24 crc kubenswrapper[4733]: I1204 19:09:24.717607 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b7wmd" event={"ID":"d4955edf-ded2-45c8-9207-c319e9356683","Type":"ContainerStarted","Data":"9a7cd701ac5df1728cf1d71a0b32829a3e98c21eddee290e459a0964c4f9cf21"} Dec 04 19:09:24 crc kubenswrapper[4733]: I1204 19:09:24.738404 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b7wmd" podStartSLOduration=2.296499433 podStartE2EDuration="4.738385106s" podCreationTimestamp="2025-12-04 19:09:20 +0000 UTC" firstStartedPulling="2025-12-04 19:09:21.691590318 +0000 UTC m=+5423.646951394" lastFinishedPulling="2025-12-04 19:09:24.133476021 +0000 UTC m=+5426.088837067" observedRunningTime="2025-12-04 19:09:24.735143629 +0000 UTC m=+5426.690504675" watchObservedRunningTime="2025-12-04 19:09:24.738385106 +0000 UTC m=+5426.693746152" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.655126 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.656664 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.660588 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.661889 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.665130 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.668737 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-nhkmm" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.677656 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.679267 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.689254 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.691416 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.701011 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.714071 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.851657 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.853038 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.855711 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-w4s2f" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.855733 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856297 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-config\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856353 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18de617e-b3eb-4a20-a530-14476a9de2e4-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856379 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fe0b9b32-c898-4636-a895-1d2a46ef40cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe0b9b32-c898-4636-a895-1d2a46ef40cc\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856400 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856418 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18de617e-b3eb-4a20-a530-14476a9de2e4-config\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856435 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856454 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18de617e-b3eb-4a20-a530-14476a9de2e4-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856470 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fd74cf23-d906-42aa-b7a3-25e9b8bbd3ff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd74cf23-d906-42aa-b7a3-25e9b8bbd3ff\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856491 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcwm2\" (UniqueName: \"kubernetes.io/projected/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-kube-api-access-kcwm2\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856517 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18de617e-b3eb-4a20-a530-14476a9de2e4-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856832 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856866 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856905 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-29985085-cb70-4280-b5f0-f8f194068f2b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29985085-cb70-4280-b5f0-f8f194068f2b\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856956 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl4fp\" (UniqueName: \"kubernetes.io/projected/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-kube-api-access-sl4fp\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.856989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-config\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.857039 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.857057 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qksnz\" (UniqueName: \"kubernetes.io/projected/18de617e-b3eb-4a20-a530-14476a9de2e4-kube-api-access-qksnz\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.857073 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.860199 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.862276 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.868669 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.870046 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.874151 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.875853 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.881269 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.900764 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.958604 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18de617e-b3eb-4a20-a530-14476a9de2e4-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959213 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fe73daf3-4a24-4813-b103-2b833c958140-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959251 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fe0b9b32-c898-4636-a895-1d2a46ef40cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe0b9b32-c898-4636-a895-1d2a46ef40cc\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959277 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dc290916-3420-463b-a53f-e44f21a20bc2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc290916-3420-463b-a53f-e44f21a20bc2\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959306 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959334 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18de617e-b3eb-4a20-a530-14476a9de2e4-config\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959360 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959383 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlvzs\" (UniqueName: \"kubernetes.io/projected/fe73daf3-4a24-4813-b103-2b833c958140-kube-api-access-mlvzs\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959405 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-config\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959423 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18de617e-b3eb-4a20-a530-14476a9de2e4-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959442 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fd74cf23-d906-42aa-b7a3-25e9b8bbd3ff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd74cf23-d906-42aa-b7a3-25e9b8bbd3ff\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959471 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcwm2\" (UniqueName: \"kubernetes.io/projected/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-kube-api-access-kcwm2\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959493 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe73daf3-4a24-4813-b103-2b833c958140-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959527 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18de617e-b3eb-4a20-a530-14476a9de2e4-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959556 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe73daf3-4a24-4813-b103-2b833c958140-config\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959583 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959614 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959635 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959662 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ccea13bb-6b95-49c7-9af0-de792de4da04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ccea13bb-6b95-49c7-9af0-de792de4da04\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959703 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr677\" (UniqueName: \"kubernetes.io/projected/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-kube-api-access-mr677\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959733 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-29985085-cb70-4280-b5f0-f8f194068f2b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29985085-cb70-4280-b5f0-f8f194068f2b\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959772 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl4fp\" (UniqueName: \"kubernetes.io/projected/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-kube-api-access-sl4fp\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959813 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe73daf3-4a24-4813-b103-2b833c958140-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959851 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-config\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959901 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959922 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qksnz\" (UniqueName: \"kubernetes.io/projected/18de617e-b3eb-4a20-a530-14476a9de2e4-kube-api-access-qksnz\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959945 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959970 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.959993 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.960020 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-config\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.961156 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18de617e-b3eb-4a20-a530-14476a9de2e4-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.961926 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.962676 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18de617e-b3eb-4a20-a530-14476a9de2e4-config\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.963031 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.964309 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-config\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.965023 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18de617e-b3eb-4a20-a530-14476a9de2e4-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.967375 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.968201 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.968229 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.968241 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fd74cf23-d906-42aa-b7a3-25e9b8bbd3ff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd74cf23-d906-42aa-b7a3-25e9b8bbd3ff\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/988b92fd029a1412dca6147f9ba6b82c93d9a5aeb14d0489a789a5d643742ce2/globalmount\"" pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.968262 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-29985085-cb70-4280-b5f0-f8f194068f2b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29985085-cb70-4280-b5f0-f8f194068f2b\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/021bc1773462bda93cec5b7db36b5bd0ee704d64214f4e1f1303d141803d745b/globalmount\"" pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.968649 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.968763 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fe0b9b32-c898-4636-a895-1d2a46ef40cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe0b9b32-c898-4636-a895-1d2a46ef40cc\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/182715a187265dc1f5eb40abd183bb6e953fa49717034f42cc128c3dbd19f90f/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.969320 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.969827 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-config\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.970406 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18de617e-b3eb-4a20-a530-14476a9de2e4-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.971480 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.973543 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.983192 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcwm2\" (UniqueName: \"kubernetes.io/projected/c88aa772-44c8-469c-a29f-eb9d3ae11b9b-kube-api-access-kcwm2\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.986923 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl4fp\" (UniqueName: \"kubernetes.io/projected/bd4e8124-0462-4c77-bc59-fb83e7ee2b25-kube-api-access-sl4fp\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:25 crc kubenswrapper[4733]: I1204 19:09:25.988305 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qksnz\" (UniqueName: \"kubernetes.io/projected/18de617e-b3eb-4a20-a530-14476a9de2e4-kube-api-access-qksnz\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.007833 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-29985085-cb70-4280-b5f0-f8f194068f2b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29985085-cb70-4280-b5f0-f8f194068f2b\") pod \"ovsdbserver-sb-2\" (UID: \"bd4e8124-0462-4c77-bc59-fb83e7ee2b25\") " pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.016458 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fd74cf23-d906-42aa-b7a3-25e9b8bbd3ff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd74cf23-d906-42aa-b7a3-25e9b8bbd3ff\") pod \"ovsdbserver-sb-1\" (UID: \"18de617e-b3eb-4a20-a530-14476a9de2e4\") " pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.021093 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fe0b9b32-c898-4636-a895-1d2a46ef40cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe0b9b32-c898-4636-a895-1d2a46ef40cc\") pod \"ovsdbserver-sb-0\" (UID: \"c88aa772-44c8-469c-a29f-eb9d3ae11b9b\") " pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.039906 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.051765 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.061855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.062291 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.062344 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.062362 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.062386 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ccea13bb-6b95-49c7-9af0-de792de4da04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ccea13bb-6b95-49c7-9af0-de792de4da04\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.062409 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr677\" (UniqueName: \"kubernetes.io/projected/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-kube-api-access-mr677\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.062585 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe73daf3-4a24-4813-b103-2b833c958140-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.063860 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27sqx\" (UniqueName: \"kubernetes.io/projected/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-kube-api-access-27sqx\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.064030 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.064131 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.064251 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fe73daf3-4a24-4813-b103-2b833c958140-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.068903 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dc290916-3420-463b-a53f-e44f21a20bc2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc290916-3420-463b-a53f-e44f21a20bc2\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.069325 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-config\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.065298 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fe73daf3-4a24-4813-b103-2b833c958140-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.065338 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.069453 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlvzs\" (UniqueName: \"kubernetes.io/projected/fe73daf3-4a24-4813-b103-2b833c958140-kube-api-access-mlvzs\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.069560 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-config\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.070516 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-config\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.069589 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-30ce8ae5-c202-4b2b-86c0-205ab0a77832\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-30ce8ae5-c202-4b2b-86c0-205ab0a77832\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.070842 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.070954 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe73daf3-4a24-4813-b103-2b833c958140-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.071504 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.071544 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ccea13bb-6b95-49c7-9af0-de792de4da04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ccea13bb-6b95-49c7-9af0-de792de4da04\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1692fda11c724b0c88ac583b71189af43db9e492be59ae84578c50161f3cab95/globalmount\"" pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.071555 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.071586 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dc290916-3420-463b-a53f-e44f21a20bc2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc290916-3420-463b-a53f-e44f21a20bc2\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bfb0210ea0bda2557cfaac017fac2f7af2886b14b3e22f07e36d4843b2b163d7/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.072025 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe73daf3-4a24-4813-b103-2b833c958140-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.072075 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe73daf3-4a24-4813-b103-2b833c958140-config\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.072115 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.072462 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe73daf3-4a24-4813-b103-2b833c958140-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.073409 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.079369 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr677\" (UniqueName: \"kubernetes.io/projected/fd6238cc-9a10-42fb-bd5c-f22151dfcd87-kube-api-access-mr677\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.083530 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe73daf3-4a24-4813-b103-2b833c958140-config\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.098959 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlvzs\" (UniqueName: \"kubernetes.io/projected/fe73daf3-4a24-4813-b103-2b833c958140-kube-api-access-mlvzs\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.109110 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ccea13bb-6b95-49c7-9af0-de792de4da04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ccea13bb-6b95-49c7-9af0-de792de4da04\") pod \"ovsdbserver-nb-2\" (UID: \"fe73daf3-4a24-4813-b103-2b833c958140\") " pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.115765 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dc290916-3420-463b-a53f-e44f21a20bc2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc290916-3420-463b-a53f-e44f21a20bc2\") pod \"ovsdbserver-nb-0\" (UID: \"fd6238cc-9a10-42fb-bd5c-f22151dfcd87\") " pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.174040 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.174392 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.174416 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.174615 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27sqx\" (UniqueName: \"kubernetes.io/projected/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-kube-api-access-27sqx\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.175500 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.175575 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.176202 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.177025 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-config\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.177093 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-30ce8ae5-c202-4b2b-86c0-205ab0a77832\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-30ce8ae5-c202-4b2b-86c0-205ab0a77832\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.177717 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-config\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.178660 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.180342 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.180374 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-30ce8ae5-c202-4b2b-86c0-205ab0a77832\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-30ce8ae5-c202-4b2b-86c0-205ab0a77832\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a8822eda0219f658c40d4676f11033aee7752610115e0595a4b58b7d5f931577/globalmount\"" pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.191434 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.206862 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27sqx\" (UniqueName: \"kubernetes.io/projected/fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3-kube-api-access-27sqx\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.241445 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-30ce8ae5-c202-4b2b-86c0-205ab0a77832\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-30ce8ae5-c202-4b2b-86c0-205ab0a77832\") pod \"ovsdbserver-nb-1\" (UID: \"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3\") " pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.501419 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.571510 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.665776 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 04 19:09:26 crc kubenswrapper[4733]: W1204 19:09:26.676777 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd4e8124_0462_4c77_bc59_fb83e7ee2b25.slice/crio-409c2d14dd8a5905743a9a2d0d42dd0c17812b1e8e0a97d1f67ba85697dcaf75 WatchSource:0}: Error finding container 409c2d14dd8a5905743a9a2d0d42dd0c17812b1e8e0a97d1f67ba85697dcaf75: Status 404 returned error can't find the container with id 409c2d14dd8a5905743a9a2d0d42dd0c17812b1e8e0a97d1f67ba85697dcaf75 Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.799383 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"bd4e8124-0462-4c77-bc59-fb83e7ee2b25","Type":"ContainerStarted","Data":"409c2d14dd8a5905743a9a2d0d42dd0c17812b1e8e0a97d1f67ba85697dcaf75"} Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.807284 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c88aa772-44c8-469c-a29f-eb9d3ae11b9b","Type":"ContainerStarted","Data":"158cc4a61dccd3bf1ef3f66daf9ed55855a75af022c2a6e67adff8eb1248b667"} Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.814436 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 04 19:09:26 crc kubenswrapper[4733]: W1204 19:09:26.826393 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe73daf3_4a24_4813_b103_2b833c958140.slice/crio-67cebe249f96c6612aa2fddcc6740efa214530126d37e850a12ff03a7b386e90 WatchSource:0}: Error finding container 67cebe249f96c6612aa2fddcc6740efa214530126d37e850a12ff03a7b386e90: Status 404 returned error can't find the container with id 67cebe249f96c6612aa2fddcc6740efa214530126d37e850a12ff03a7b386e90 Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.874224 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 19:09:26 crc kubenswrapper[4733]: W1204 19:09:26.880632 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd6238cc_9a10_42fb_bd5c_f22151dfcd87.slice/crio-34087c13971aacf583a095c5315a17654a8243aa8b9d1dbcf3e693ff8275e1fd WatchSource:0}: Error finding container 34087c13971aacf583a095c5315a17654a8243aa8b9d1dbcf3e693ff8275e1fd: Status 404 returned error can't find the container with id 34087c13971aacf583a095c5315a17654a8243aa8b9d1dbcf3e693ff8275e1fd Dec 04 19:09:26 crc kubenswrapper[4733]: I1204 19:09:26.957832 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 04 19:09:26 crc kubenswrapper[4733]: W1204 19:09:26.972544 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb48f9f0_db9c_4f91_801b_e80f1bd2f7c3.slice/crio-dfa6c1cacaad9aee5c090afbec40395dccc5eb59f1ee8b4f9de77d0f7f9c7da9 WatchSource:0}: Error finding container dfa6c1cacaad9aee5c090afbec40395dccc5eb59f1ee8b4f9de77d0f7f9c7da9: Status 404 returned error can't find the container with id dfa6c1cacaad9aee5c090afbec40395dccc5eb59f1ee8b4f9de77d0f7f9c7da9 Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.392740 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.819554 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c88aa772-44c8-469c-a29f-eb9d3ae11b9b","Type":"ContainerStarted","Data":"a6d64511915815e783cf3c23eadea6ffac468a90b3bba6091f7200edb95f7203"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.820235 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c88aa772-44c8-469c-a29f-eb9d3ae11b9b","Type":"ContainerStarted","Data":"4ce4a3705a1ae41ac63c101f40e0629defaeb29467e23c80bf071e7936aaa3dd"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.822994 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"18de617e-b3eb-4a20-a530-14476a9de2e4","Type":"ContainerStarted","Data":"b3b30b167a1ef63a1966dc889c8681263bc9e8f8d4cbf32ff576f13bbd559597"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.823054 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"18de617e-b3eb-4a20-a530-14476a9de2e4","Type":"ContainerStarted","Data":"e713892d6ee7a21c5dba1f776ad2df729414cd902a5c1ec2671ed2c5e518ccd6"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.830791 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"fd6238cc-9a10-42fb-bd5c-f22151dfcd87","Type":"ContainerStarted","Data":"9e22cfad8b12b22021056984df96bd084ddd6a6d886c50cc61aa333d0eb14bc6"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.830861 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"fd6238cc-9a10-42fb-bd5c-f22151dfcd87","Type":"ContainerStarted","Data":"838ac9f809542f3aaef7722d6a7f6787dc83f697f0031f22927f003a66afa8a9"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.830875 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"fd6238cc-9a10-42fb-bd5c-f22151dfcd87","Type":"ContainerStarted","Data":"34087c13971aacf583a095c5315a17654a8243aa8b9d1dbcf3e693ff8275e1fd"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.833844 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3","Type":"ContainerStarted","Data":"4eebbe84a34d1a418cbc85aeca89b1a2173268532b16e22a25e4b26a2ac6f2c4"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.833944 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3","Type":"ContainerStarted","Data":"94fdce8068c8e094663091f7a92c38764b7b7c5e47d884a8ced99f5f2d98b3df"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.833966 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3","Type":"ContainerStarted","Data":"dfa6c1cacaad9aee5c090afbec40395dccc5eb59f1ee8b4f9de77d0f7f9c7da9"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.841445 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"bd4e8124-0462-4c77-bc59-fb83e7ee2b25","Type":"ContainerStarted","Data":"e19dc742dff4ca0ae5083704b1f2cb43874425be34b435ddf931687d2ad0ac4a"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.841596 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"bd4e8124-0462-4c77-bc59-fb83e7ee2b25","Type":"ContainerStarted","Data":"d968b7d448035b02cbe4e6f04be2666684c3c2ec36fb8e9614222cce82ef05d2"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.846905 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"fe73daf3-4a24-4813-b103-2b833c958140","Type":"ContainerStarted","Data":"1a814ad0eb972335d4791f948552993cc013d127cbb7d86f41aaf2bf103f6a17"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.846949 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"fe73daf3-4a24-4813-b103-2b833c958140","Type":"ContainerStarted","Data":"d765fe77c192ca4040531b7ca54c8878f5878782b7e7e43eb871648e2ed15bfd"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.846961 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"fe73daf3-4a24-4813-b103-2b833c958140","Type":"ContainerStarted","Data":"67cebe249f96c6612aa2fddcc6740efa214530126d37e850a12ff03a7b386e90"} Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.850666 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.850648745 podStartE2EDuration="3.850648745s" podCreationTimestamp="2025-12-04 19:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:27.840057048 +0000 UTC m=+5429.795418114" watchObservedRunningTime="2025-12-04 19:09:27.850648745 +0000 UTC m=+5429.806009791" Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.859550 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.859530045 podStartE2EDuration="3.859530045s" podCreationTimestamp="2025-12-04 19:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:27.858766544 +0000 UTC m=+5429.814127600" watchObservedRunningTime="2025-12-04 19:09:27.859530045 +0000 UTC m=+5429.814891091" Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.880037 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.8800184890000002 podStartE2EDuration="3.880018489s" podCreationTimestamp="2025-12-04 19:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:27.87821304 +0000 UTC m=+5429.833574086" watchObservedRunningTime="2025-12-04 19:09:27.880018489 +0000 UTC m=+5429.835379525" Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.903689 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.903670308 podStartE2EDuration="3.903670308s" podCreationTimestamp="2025-12-04 19:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:27.900000639 +0000 UTC m=+5429.855361685" watchObservedRunningTime="2025-12-04 19:09:27.903670308 +0000 UTC m=+5429.859031354" Dec 04 19:09:27 crc kubenswrapper[4733]: I1204 19:09:27.918814 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.918776476 podStartE2EDuration="3.918776476s" podCreationTimestamp="2025-12-04 19:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:27.915722654 +0000 UTC m=+5429.871083710" watchObservedRunningTime="2025-12-04 19:09:27.918776476 +0000 UTC m=+5429.874137522" Dec 04 19:09:28 crc kubenswrapper[4733]: I1204 19:09:28.859628 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"18de617e-b3eb-4a20-a530-14476a9de2e4","Type":"ContainerStarted","Data":"99dea8507136909118ac2634e9724122b38a6638432368fe5b780d97b6988b00"} Dec 04 19:09:28 crc kubenswrapper[4733]: I1204 19:09:28.895414 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.895389732 podStartE2EDuration="4.895389732s" podCreationTimestamp="2025-12-04 19:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:28.883201833 +0000 UTC m=+5430.838562929" watchObservedRunningTime="2025-12-04 19:09:28.895389732 +0000 UTC m=+5430.850750818" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.041106 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.052365 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.062418 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.087934 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.098700 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.176027 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.192154 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.501954 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.871955 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:29 crc kubenswrapper[4733]: I1204 19:09:29.871989 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:30 crc kubenswrapper[4733]: I1204 19:09:30.838725 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:30 crc kubenswrapper[4733]: I1204 19:09:30.838780 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:30 crc kubenswrapper[4733]: I1204 19:09:30.905643 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:30 crc kubenswrapper[4733]: E1204 19:09:30.912982 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice/crio-783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice\": RecentStats: unable to find data in memory cache]" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.062242 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.081831 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.089915 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.176598 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.192375 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.430645 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76c96887b9-fnzdj"] Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.445731 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.449583 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c96887b9-fnzdj"] Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.465188 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.505246 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.587357 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-dns-svc\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.587415 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcdsn\" (UniqueName: \"kubernetes.io/projected/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-kube-api-access-tcdsn\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.587464 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.587536 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-ovsdbserver-sb\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.689015 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-ovsdbserver-sb\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.689098 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-dns-svc\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.689133 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcdsn\" (UniqueName: \"kubernetes.io/projected/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-kube-api-access-tcdsn\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.689164 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.690051 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.690453 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-dns-svc\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.690716 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-ovsdbserver-sb\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.711864 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcdsn\" (UniqueName: \"kubernetes.io/projected/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-kube-api-access-tcdsn\") pod \"dnsmasq-dns-76c96887b9-fnzdj\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.822349 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:31 crc kubenswrapper[4733]: I1204 19:09:31.997573 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.107213 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.219555 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.265889 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.303330 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.315537 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c96887b9-fnzdj"] Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.316684 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.546125 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.596019 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.596352 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c96887b9-fnzdj"] Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.619472 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66775b999f-sdbwg"] Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.620814 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.624182 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.664441 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66775b999f-sdbwg"] Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.711701 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-dns-svc\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.711919 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-sb\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.711955 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-nb\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.711977 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-config\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.711996 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj5lt\" (UniqueName: \"kubernetes.io/projected/4ff0d2bd-c12a-4ebd-8649-3954979678b3-kube-api-access-zj5lt\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.813696 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-nb\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.813742 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-config\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.813765 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj5lt\" (UniqueName: \"kubernetes.io/projected/4ff0d2bd-c12a-4ebd-8649-3954979678b3-kube-api-access-zj5lt\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.813882 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-dns-svc\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.813900 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-sb\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.814718 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-sb\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.814716 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-nb\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.814844 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-config\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.814874 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-dns-svc\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.834212 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj5lt\" (UniqueName: \"kubernetes.io/projected/4ff0d2bd-c12a-4ebd-8649-3954979678b3-kube-api-access-zj5lt\") pod \"dnsmasq-dns-66775b999f-sdbwg\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.894212 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ee10692-3ca8-4b91-9188-9cc8f52b8a80" containerID="f0e0b357be1c92bb330ddcd76fd5872ccde4ac13d698643af67492a757514d1c" exitCode=0 Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.895079 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" event={"ID":"8ee10692-3ca8-4b91-9188-9cc8f52b8a80","Type":"ContainerDied","Data":"f0e0b357be1c92bb330ddcd76fd5872ccde4ac13d698643af67492a757514d1c"} Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.895110 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" event={"ID":"8ee10692-3ca8-4b91-9188-9cc8f52b8a80","Type":"ContainerStarted","Data":"edc88f3c17b2e3fe11125a282e00ecdc5ae5491e92b0be767a5723c35450f948"} Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.942420 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:32 crc kubenswrapper[4733]: I1204 19:09:32.950378 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.313355 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.428534 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config\") pod \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.428605 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcdsn\" (UniqueName: \"kubernetes.io/projected/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-kube-api-access-tcdsn\") pod \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.428688 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-ovsdbserver-sb\") pod \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.428737 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-dns-svc\") pod \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.437227 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-kube-api-access-tcdsn" (OuterVolumeSpecName: "kube-api-access-tcdsn") pod "8ee10692-3ca8-4b91-9188-9cc8f52b8a80" (UID: "8ee10692-3ca8-4b91-9188-9cc8f52b8a80"). InnerVolumeSpecName "kube-api-access-tcdsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.449869 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8ee10692-3ca8-4b91-9188-9cc8f52b8a80" (UID: "8ee10692-3ca8-4b91-9188-9cc8f52b8a80"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:09:33 crc kubenswrapper[4733]: E1204 19:09:33.456206 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config podName:8ee10692-3ca8-4b91-9188-9cc8f52b8a80 nodeName:}" failed. No retries permitted until 2025-12-04 19:09:33.956174086 +0000 UTC m=+5435.911535132 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config") pod "8ee10692-3ca8-4b91-9188-9cc8f52b8a80" (UID: "8ee10692-3ca8-4b91-9188-9cc8f52b8a80") : error deleting /var/lib/kubelet/pods/8ee10692-3ca8-4b91-9188-9cc8f52b8a80/volume-subpaths: remove /var/lib/kubelet/pods/8ee10692-3ca8-4b91-9188-9cc8f52b8a80/volume-subpaths: no such file or directory Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.456442 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ee10692-3ca8-4b91-9188-9cc8f52b8a80" (UID: "8ee10692-3ca8-4b91-9188-9cc8f52b8a80"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.525360 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66775b999f-sdbwg"] Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.532691 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcdsn\" (UniqueName: \"kubernetes.io/projected/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-kube-api-access-tcdsn\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.532741 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.532757 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:33 crc kubenswrapper[4733]: W1204 19:09:33.538918 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ff0d2bd_c12a_4ebd_8649_3954979678b3.slice/crio-eb822f5d5385ddace766897adf8751abac5ccb47cec3026efbdcd96872069dc0 WatchSource:0}: Error finding container eb822f5d5385ddace766897adf8751abac5ccb47cec3026efbdcd96872069dc0: Status 404 returned error can't find the container with id eb822f5d5385ddace766897adf8751abac5ccb47cec3026efbdcd96872069dc0 Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.907909 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" event={"ID":"8ee10692-3ca8-4b91-9188-9cc8f52b8a80","Type":"ContainerDied","Data":"edc88f3c17b2e3fe11125a282e00ecdc5ae5491e92b0be767a5723c35450f948"} Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.907970 4733 scope.go:117] "RemoveContainer" containerID="f0e0b357be1c92bb330ddcd76fd5872ccde4ac13d698643af67492a757514d1c" Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.907984 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c96887b9-fnzdj" Dec 04 19:09:33 crc kubenswrapper[4733]: I1204 19:09:33.909772 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" event={"ID":"4ff0d2bd-c12a-4ebd-8649-3954979678b3","Type":"ContainerStarted","Data":"eb822f5d5385ddace766897adf8751abac5ccb47cec3026efbdcd96872069dc0"} Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.043158 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config\") pod \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\" (UID: \"8ee10692-3ca8-4b91-9188-9cc8f52b8a80\") " Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.045578 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config" (OuterVolumeSpecName: "config") pod "8ee10692-3ca8-4b91-9188-9cc8f52b8a80" (UID: "8ee10692-3ca8-4b91-9188-9cc8f52b8a80"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.146406 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ee10692-3ca8-4b91-9188-9cc8f52b8a80-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.298815 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c96887b9-fnzdj"] Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.308722 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76c96887b9-fnzdj"] Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.349859 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ee10692-3ca8-4b91-9188-9cc8f52b8a80" path="/var/lib/kubelet/pods/8ee10692-3ca8-4b91-9188-9cc8f52b8a80/volumes" Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.683774 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b7wmd"] Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.684206 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b7wmd" podUID="d4955edf-ded2-45c8-9207-c319e9356683" containerName="registry-server" containerID="cri-o://9a7cd701ac5df1728cf1d71a0b32829a3e98c21eddee290e459a0964c4f9cf21" gracePeriod=2 Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.935783 4733 generic.go:334] "Generic (PLEG): container finished" podID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" containerID="b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656" exitCode=0 Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.935870 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" event={"ID":"4ff0d2bd-c12a-4ebd-8649-3954979678b3","Type":"ContainerDied","Data":"b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656"} Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.940635 4733 generic.go:334] "Generic (PLEG): container finished" podID="d4955edf-ded2-45c8-9207-c319e9356683" containerID="9a7cd701ac5df1728cf1d71a0b32829a3e98c21eddee290e459a0964c4f9cf21" exitCode=0 Dec 04 19:09:34 crc kubenswrapper[4733]: I1204 19:09:34.940672 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b7wmd" event={"ID":"d4955edf-ded2-45c8-9207-c319e9356683","Type":"ContainerDied","Data":"9a7cd701ac5df1728cf1d71a0b32829a3e98c21eddee290e459a0964c4f9cf21"} Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.122781 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.171567 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9vln\" (UniqueName: \"kubernetes.io/projected/d4955edf-ded2-45c8-9207-c319e9356683-kube-api-access-q9vln\") pod \"d4955edf-ded2-45c8-9207-c319e9356683\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.171626 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-catalog-content\") pod \"d4955edf-ded2-45c8-9207-c319e9356683\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.171729 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-utilities\") pod \"d4955edf-ded2-45c8-9207-c319e9356683\" (UID: \"d4955edf-ded2-45c8-9207-c319e9356683\") " Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.172813 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-utilities" (OuterVolumeSpecName: "utilities") pod "d4955edf-ded2-45c8-9207-c319e9356683" (UID: "d4955edf-ded2-45c8-9207-c319e9356683"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.175900 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4955edf-ded2-45c8-9207-c319e9356683-kube-api-access-q9vln" (OuterVolumeSpecName: "kube-api-access-q9vln") pod "d4955edf-ded2-45c8-9207-c319e9356683" (UID: "d4955edf-ded2-45c8-9207-c319e9356683"). InnerVolumeSpecName "kube-api-access-q9vln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.185813 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.185856 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9vln\" (UniqueName: \"kubernetes.io/projected/d4955edf-ded2-45c8-9207-c319e9356683-kube-api-access-q9vln\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.223925 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4955edf-ded2-45c8-9207-c319e9356683" (UID: "d4955edf-ded2-45c8-9207-c319e9356683"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.287931 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4955edf-ded2-45c8-9207-c319e9356683-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.640261 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Dec 04 19:09:35 crc kubenswrapper[4733]: E1204 19:09:35.640710 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4955edf-ded2-45c8-9207-c319e9356683" containerName="extract-utilities" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.640729 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4955edf-ded2-45c8-9207-c319e9356683" containerName="extract-utilities" Dec 04 19:09:35 crc kubenswrapper[4733]: E1204 19:09:35.640755 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee10692-3ca8-4b91-9188-9cc8f52b8a80" containerName="init" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.640766 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee10692-3ca8-4b91-9188-9cc8f52b8a80" containerName="init" Dec 04 19:09:35 crc kubenswrapper[4733]: E1204 19:09:35.640836 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4955edf-ded2-45c8-9207-c319e9356683" containerName="extract-content" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.640850 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4955edf-ded2-45c8-9207-c319e9356683" containerName="extract-content" Dec 04 19:09:35 crc kubenswrapper[4733]: E1204 19:09:35.640868 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4955edf-ded2-45c8-9207-c319e9356683" containerName="registry-server" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.640878 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4955edf-ded2-45c8-9207-c319e9356683" containerName="registry-server" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.643186 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ee10692-3ca8-4b91-9188-9cc8f52b8a80" containerName="init" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.643624 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4955edf-ded2-45c8-9207-c319e9356683" containerName="registry-server" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.644598 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.649042 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.665889 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.694285 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtdpf\" (UniqueName: \"kubernetes.io/projected/0c73247a-c966-401a-944d-31c43164b6b7-kube-api-access-xtdpf\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.694330 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0c73247a-c966-401a-944d-31c43164b6b7-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.694424 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.796601 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.797217 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtdpf\" (UniqueName: \"kubernetes.io/projected/0c73247a-c966-401a-944d-31c43164b6b7-kube-api-access-xtdpf\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.797452 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0c73247a-c966-401a-944d-31c43164b6b7-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.804306 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.804356 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/585a2ff467665e019c84976e43780486ab2bdf2b1c52d53a2844085c000ba112/globalmount\"" pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.804657 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0c73247a-c966-401a-944d-31c43164b6b7-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.823888 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtdpf\" (UniqueName: \"kubernetes.io/projected/0c73247a-c966-401a-944d-31c43164b6b7-kube-api-access-xtdpf\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.855712 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\") pod \"ovn-copy-data\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.953203 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" event={"ID":"4ff0d2bd-c12a-4ebd-8649-3954979678b3","Type":"ContainerStarted","Data":"aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87"} Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.953395 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.955962 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b7wmd" event={"ID":"d4955edf-ded2-45c8-9207-c319e9356683","Type":"ContainerDied","Data":"559d5e9a01e2d33bccc497d92f56d635ac5c615d2f7cfbe7e01b82346d8038f7"} Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.956000 4733 scope.go:117] "RemoveContainer" containerID="9a7cd701ac5df1728cf1d71a0b32829a3e98c21eddee290e459a0964c4f9cf21" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.956042 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b7wmd" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.973990 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.978225 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" podStartSLOduration=3.978203315 podStartE2EDuration="3.978203315s" podCreationTimestamp="2025-12-04 19:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:35.973292733 +0000 UTC m=+5437.928653809" watchObservedRunningTime="2025-12-04 19:09:35.978203315 +0000 UTC m=+5437.933564371" Dec 04 19:09:35 crc kubenswrapper[4733]: I1204 19:09:35.987091 4733 scope.go:117] "RemoveContainer" containerID="ed31019c4114fdafcdab7e1831696b2e6b08dc2c00e913c1eee148e5a17ffd6d" Dec 04 19:09:36 crc kubenswrapper[4733]: I1204 19:09:36.018459 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b7wmd"] Dec 04 19:09:36 crc kubenswrapper[4733]: I1204 19:09:36.024695 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b7wmd"] Dec 04 19:09:36 crc kubenswrapper[4733]: I1204 19:09:36.041214 4733 scope.go:117] "RemoveContainer" containerID="81f3e36d6e0ad74c69e6a565d010513a0ab5f37fc445e0fddf5c6086321d4918" Dec 04 19:09:36 crc kubenswrapper[4733]: I1204 19:09:36.350052 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4955edf-ded2-45c8-9207-c319e9356683" path="/var/lib/kubelet/pods/d4955edf-ded2-45c8-9207-c319e9356683/volumes" Dec 04 19:09:36 crc kubenswrapper[4733]: I1204 19:09:36.566405 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 19:09:36 crc kubenswrapper[4733]: W1204 19:09:36.568670 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c73247a_c966_401a_944d_31c43164b6b7.slice/crio-48f7f1a986182b71bc0535f509fca3d9f8cbb17c281cd0cd8179a54ab9e6bbe8 WatchSource:0}: Error finding container 48f7f1a986182b71bc0535f509fca3d9f8cbb17c281cd0cd8179a54ab9e6bbe8: Status 404 returned error can't find the container with id 48f7f1a986182b71bc0535f509fca3d9f8cbb17c281cd0cd8179a54ab9e6bbe8 Dec 04 19:09:36 crc kubenswrapper[4733]: I1204 19:09:36.965826 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"0c73247a-c966-401a-944d-31c43164b6b7","Type":"ContainerStarted","Data":"48f7f1a986182b71bc0535f509fca3d9f8cbb17c281cd0cd8179a54ab9e6bbe8"} Dec 04 19:09:37 crc kubenswrapper[4733]: I1204 19:09:37.980382 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"0c73247a-c966-401a-944d-31c43164b6b7","Type":"ContainerStarted","Data":"b259aacc064a590c59a956ec0d0863a146ed74215a33276852ca9599d0d3e41b"} Dec 04 19:09:38 crc kubenswrapper[4733]: I1204 19:09:38.012641 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.35571602 podStartE2EDuration="4.012614761s" podCreationTimestamp="2025-12-04 19:09:34 +0000 UTC" firstStartedPulling="2025-12-04 19:09:36.572336199 +0000 UTC m=+5438.527697285" lastFinishedPulling="2025-12-04 19:09:37.22923497 +0000 UTC m=+5439.184596026" observedRunningTime="2025-12-04 19:09:38.002229779 +0000 UTC m=+5439.957590885" watchObservedRunningTime="2025-12-04 19:09:38.012614761 +0000 UTC m=+5439.967975837" Dec 04 19:09:41 crc kubenswrapper[4733]: E1204 19:09:41.131509 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice/crio-783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice\": RecentStats: unable to find data in memory cache]" Dec 04 19:09:42 crc kubenswrapper[4733]: I1204 19:09:42.945049 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.033423 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f79bf7859-x6fqr"] Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.033668 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" podUID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" containerName="dnsmasq-dns" containerID="cri-o://ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49" gracePeriod=10 Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.527918 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.639935 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-dns-svc\") pod \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.639994 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw2tf\" (UniqueName: \"kubernetes.io/projected/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-kube-api-access-zw2tf\") pod \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.640165 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-config\") pod \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\" (UID: \"2d3818e0-89ec-4cfd-8bd1-918c99424ae0\") " Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.655394 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-kube-api-access-zw2tf" (OuterVolumeSpecName: "kube-api-access-zw2tf") pod "2d3818e0-89ec-4cfd-8bd1-918c99424ae0" (UID: "2d3818e0-89ec-4cfd-8bd1-918c99424ae0"). InnerVolumeSpecName "kube-api-access-zw2tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.675815 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-config" (OuterVolumeSpecName: "config") pod "2d3818e0-89ec-4cfd-8bd1-918c99424ae0" (UID: "2d3818e0-89ec-4cfd-8bd1-918c99424ae0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.675814 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d3818e0-89ec-4cfd-8bd1-918c99424ae0" (UID: "2d3818e0-89ec-4cfd-8bd1-918c99424ae0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.741867 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.741897 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw2tf\" (UniqueName: \"kubernetes.io/projected/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-kube-api-access-zw2tf\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:43 crc kubenswrapper[4733]: I1204 19:09:43.741906 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d3818e0-89ec-4cfd-8bd1-918c99424ae0-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.059572 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.059508 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" containerID="ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49" exitCode=0 Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.059596 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" event={"ID":"2d3818e0-89ec-4cfd-8bd1-918c99424ae0","Type":"ContainerDied","Data":"ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49"} Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.059858 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f79bf7859-x6fqr" event={"ID":"2d3818e0-89ec-4cfd-8bd1-918c99424ae0","Type":"ContainerDied","Data":"2579a7d57539774294db6414d653b1ce2d17db92b523234a7ecbe697745a794a"} Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.059937 4733 scope.go:117] "RemoveContainer" containerID="ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.091286 4733 scope.go:117] "RemoveContainer" containerID="938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.116936 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f79bf7859-x6fqr"] Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.123062 4733 scope.go:117] "RemoveContainer" containerID="ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49" Dec 04 19:09:44 crc kubenswrapper[4733]: E1204 19:09:44.123633 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49\": container with ID starting with ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49 not found: ID does not exist" containerID="ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.123697 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49"} err="failed to get container status \"ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49\": rpc error: code = NotFound desc = could not find container \"ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49\": container with ID starting with ae1a1571c57a53aafc758e596600f027e63f9b69bc86d08bb59c1361eef12c49 not found: ID does not exist" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.123739 4733 scope.go:117] "RemoveContainer" containerID="938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59" Dec 04 19:09:44 crc kubenswrapper[4733]: E1204 19:09:44.124964 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59\": container with ID starting with 938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59 not found: ID does not exist" containerID="938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.125046 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f79bf7859-x6fqr"] Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.125038 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59"} err="failed to get container status \"938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59\": rpc error: code = NotFound desc = could not find container \"938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59\": container with ID starting with 938ab31f6681aeb0371a505950bb3bc269db68353d06ea006e148d02ce49ba59 not found: ID does not exist" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.356917 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" path="/var/lib/kubelet/pods/2d3818e0-89ec-4cfd-8bd1-918c99424ae0/volumes" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.653407 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 04 19:09:44 crc kubenswrapper[4733]: E1204 19:09:44.653931 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" containerName="dnsmasq-dns" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.653959 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" containerName="dnsmasq-dns" Dec 04 19:09:44 crc kubenswrapper[4733]: E1204 19:09:44.653998 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" containerName="init" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.654010 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" containerName="init" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.654382 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d3818e0-89ec-4cfd-8bd1-918c99424ae0" containerName="dnsmasq-dns" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.656036 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.659007 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.659080 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.659590 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.661169 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-t8tmz" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.759249 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.759342 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-config\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.759438 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.759501 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-scripts\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.759571 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc4bt\" (UniqueName: \"kubernetes.io/projected/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-kube-api-access-gc4bt\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.860697 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.860815 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-scripts\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.860889 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc4bt\" (UniqueName: \"kubernetes.io/projected/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-kube-api-access-gc4bt\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.860987 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.861046 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-config\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.861861 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-scripts\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.861958 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.862323 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-config\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.866132 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.880066 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc4bt\" (UniqueName: \"kubernetes.io/projected/5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0-kube-api-access-gc4bt\") pod \"ovn-northd-0\" (UID: \"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0\") " pod="openstack/ovn-northd-0" Dec 04 19:09:44 crc kubenswrapper[4733]: I1204 19:09:44.976031 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 19:09:45 crc kubenswrapper[4733]: I1204 19:09:45.404614 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 19:09:46 crc kubenswrapper[4733]: I1204 19:09:46.082139 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0","Type":"ContainerStarted","Data":"8cd805510df1a05ea1ca7d63c5351a6b03a2c8a0dc8348cfed5e06feff883444"} Dec 04 19:09:46 crc kubenswrapper[4733]: I1204 19:09:46.082483 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0","Type":"ContainerStarted","Data":"fcc09881cce5c1b5ffbc0d0099c20dc6355714a47bd074efb1d8c52ecf0489c2"} Dec 04 19:09:47 crc kubenswrapper[4733]: I1204 19:09:47.092604 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0","Type":"ContainerStarted","Data":"eedf09d9ed4403b7378293be2698b85caa5a09eed88f9e4a598bc52fddc74fbd"} Dec 04 19:09:47 crc kubenswrapper[4733]: I1204 19:09:47.092997 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 04 19:09:47 crc kubenswrapper[4733]: I1204 19:09:47.113901 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.113881186 podStartE2EDuration="3.113881186s" podCreationTimestamp="2025-12-04 19:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:47.111192024 +0000 UTC m=+5449.066553080" watchObservedRunningTime="2025-12-04 19:09:47.113881186 +0000 UTC m=+5449.069242232" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.207429 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bsb7q"] Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.209213 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.223971 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bsb7q"] Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.301906 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f755-account-create-update-sgvzv"] Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.302914 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.305177 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.312570 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f755-account-create-update-sgvzv"] Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.394419 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmjw9\" (UniqueName: \"kubernetes.io/projected/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-kube-api-access-fmjw9\") pod \"keystone-db-create-bsb7q\" (UID: \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\") " pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.395472 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-operator-scripts\") pod \"keystone-db-create-bsb7q\" (UID: \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\") " pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.496648 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmjw9\" (UniqueName: \"kubernetes.io/projected/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-kube-api-access-fmjw9\") pod \"keystone-db-create-bsb7q\" (UID: \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\") " pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.496748 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-operator-scripts\") pod \"keystone-db-create-bsb7q\" (UID: \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\") " pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.496825 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj86w\" (UniqueName: \"kubernetes.io/projected/e23e0f7f-8466-4896-9aae-0fd66d607599-kube-api-access-vj86w\") pod \"keystone-f755-account-create-update-sgvzv\" (UID: \"e23e0f7f-8466-4896-9aae-0fd66d607599\") " pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.496856 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23e0f7f-8466-4896-9aae-0fd66d607599-operator-scripts\") pod \"keystone-f755-account-create-update-sgvzv\" (UID: \"e23e0f7f-8466-4896-9aae-0fd66d607599\") " pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.497762 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-operator-scripts\") pod \"keystone-db-create-bsb7q\" (UID: \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\") " pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.521515 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmjw9\" (UniqueName: \"kubernetes.io/projected/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-kube-api-access-fmjw9\") pod \"keystone-db-create-bsb7q\" (UID: \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\") " pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.539395 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.598528 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj86w\" (UniqueName: \"kubernetes.io/projected/e23e0f7f-8466-4896-9aae-0fd66d607599-kube-api-access-vj86w\") pod \"keystone-f755-account-create-update-sgvzv\" (UID: \"e23e0f7f-8466-4896-9aae-0fd66d607599\") " pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.598912 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23e0f7f-8466-4896-9aae-0fd66d607599-operator-scripts\") pod \"keystone-f755-account-create-update-sgvzv\" (UID: \"e23e0f7f-8466-4896-9aae-0fd66d607599\") " pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.599671 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23e0f7f-8466-4896-9aae-0fd66d607599-operator-scripts\") pod \"keystone-f755-account-create-update-sgvzv\" (UID: \"e23e0f7f-8466-4896-9aae-0fd66d607599\") " pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.626995 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj86w\" (UniqueName: \"kubernetes.io/projected/e23e0f7f-8466-4896-9aae-0fd66d607599-kube-api-access-vj86w\") pod \"keystone-f755-account-create-update-sgvzv\" (UID: \"e23e0f7f-8466-4896-9aae-0fd66d607599\") " pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:50 crc kubenswrapper[4733]: I1204 19:09:50.917180 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:51 crc kubenswrapper[4733]: I1204 19:09:51.061876 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bsb7q"] Dec 04 19:09:51 crc kubenswrapper[4733]: W1204 19:09:51.078315 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f99dc23_04af_4f68_b2cc_7f9b5f29798d.slice/crio-63d46af368f1ddc31287cbe2422ee53b0806b4af0d5dd3f90f1352388a5790f6 WatchSource:0}: Error finding container 63d46af368f1ddc31287cbe2422ee53b0806b4af0d5dd3f90f1352388a5790f6: Status 404 returned error can't find the container with id 63d46af368f1ddc31287cbe2422ee53b0806b4af0d5dd3f90f1352388a5790f6 Dec 04 19:09:51 crc kubenswrapper[4733]: I1204 19:09:51.126706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bsb7q" event={"ID":"4f99dc23-04af-4f68-b2cc-7f9b5f29798d","Type":"ContainerStarted","Data":"63d46af368f1ddc31287cbe2422ee53b0806b4af0d5dd3f90f1352388a5790f6"} Dec 04 19:09:51 crc kubenswrapper[4733]: E1204 19:09:51.349298 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice/crio-783caae34898390c7186a21ec61ed142548141157748ce1df63617b5cbc3e3e4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf455a245_3a28_48c0_b83a_5b5b56c897f6.slice\": RecentStats: unable to find data in memory cache]" Dec 04 19:09:51 crc kubenswrapper[4733]: I1204 19:09:51.355591 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f755-account-create-update-sgvzv"] Dec 04 19:09:51 crc kubenswrapper[4733]: W1204 19:09:51.358165 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode23e0f7f_8466_4896_9aae_0fd66d607599.slice/crio-d1a54f6ec2bbfa749ed314db099976d1b4d7fec4777af230829c61b22d37b64d WatchSource:0}: Error finding container d1a54f6ec2bbfa749ed314db099976d1b4d7fec4777af230829c61b22d37b64d: Status 404 returned error can't find the container with id d1a54f6ec2bbfa749ed314db099976d1b4d7fec4777af230829c61b22d37b64d Dec 04 19:09:52 crc kubenswrapper[4733]: I1204 19:09:52.139706 4733 generic.go:334] "Generic (PLEG): container finished" podID="4f99dc23-04af-4f68-b2cc-7f9b5f29798d" containerID="cb9de33098fddc8c46c84a918c754948d9f51cac41ce8e66745ee9dbd028e6f3" exitCode=0 Dec 04 19:09:52 crc kubenswrapper[4733]: I1204 19:09:52.139772 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bsb7q" event={"ID":"4f99dc23-04af-4f68-b2cc-7f9b5f29798d","Type":"ContainerDied","Data":"cb9de33098fddc8c46c84a918c754948d9f51cac41ce8e66745ee9dbd028e6f3"} Dec 04 19:09:52 crc kubenswrapper[4733]: I1204 19:09:52.142905 4733 generic.go:334] "Generic (PLEG): container finished" podID="e23e0f7f-8466-4896-9aae-0fd66d607599" containerID="48847da5f73ece7aa31649bab70a49d9d756ea52b466ef1bfa18de1e9e821497" exitCode=0 Dec 04 19:09:52 crc kubenswrapper[4733]: I1204 19:09:52.142968 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f755-account-create-update-sgvzv" event={"ID":"e23e0f7f-8466-4896-9aae-0fd66d607599","Type":"ContainerDied","Data":"48847da5f73ece7aa31649bab70a49d9d756ea52b466ef1bfa18de1e9e821497"} Dec 04 19:09:52 crc kubenswrapper[4733]: I1204 19:09:52.143030 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f755-account-create-update-sgvzv" event={"ID":"e23e0f7f-8466-4896-9aae-0fd66d607599","Type":"ContainerStarted","Data":"d1a54f6ec2bbfa749ed314db099976d1b4d7fec4777af230829c61b22d37b64d"} Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.576078 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.659864 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.752056 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmjw9\" (UniqueName: \"kubernetes.io/projected/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-kube-api-access-fmjw9\") pod \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\" (UID: \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\") " Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.752165 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23e0f7f-8466-4896-9aae-0fd66d607599-operator-scripts\") pod \"e23e0f7f-8466-4896-9aae-0fd66d607599\" (UID: \"e23e0f7f-8466-4896-9aae-0fd66d607599\") " Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.752203 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj86w\" (UniqueName: \"kubernetes.io/projected/e23e0f7f-8466-4896-9aae-0fd66d607599-kube-api-access-vj86w\") pod \"e23e0f7f-8466-4896-9aae-0fd66d607599\" (UID: \"e23e0f7f-8466-4896-9aae-0fd66d607599\") " Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.752293 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-operator-scripts\") pod \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\" (UID: \"4f99dc23-04af-4f68-b2cc-7f9b5f29798d\") " Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.753009 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4f99dc23-04af-4f68-b2cc-7f9b5f29798d" (UID: "4f99dc23-04af-4f68-b2cc-7f9b5f29798d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.753009 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e23e0f7f-8466-4896-9aae-0fd66d607599-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e23e0f7f-8466-4896-9aae-0fd66d607599" (UID: "e23e0f7f-8466-4896-9aae-0fd66d607599"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.758053 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-kube-api-access-fmjw9" (OuterVolumeSpecName: "kube-api-access-fmjw9") pod "4f99dc23-04af-4f68-b2cc-7f9b5f29798d" (UID: "4f99dc23-04af-4f68-b2cc-7f9b5f29798d"). InnerVolumeSpecName "kube-api-access-fmjw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.758170 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23e0f7f-8466-4896-9aae-0fd66d607599-kube-api-access-vj86w" (OuterVolumeSpecName: "kube-api-access-vj86w") pod "e23e0f7f-8466-4896-9aae-0fd66d607599" (UID: "e23e0f7f-8466-4896-9aae-0fd66d607599"). InnerVolumeSpecName "kube-api-access-vj86w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.854998 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmjw9\" (UniqueName: \"kubernetes.io/projected/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-kube-api-access-fmjw9\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.855062 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e23e0f7f-8466-4896-9aae-0fd66d607599-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.855080 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj86w\" (UniqueName: \"kubernetes.io/projected/e23e0f7f-8466-4896-9aae-0fd66d607599-kube-api-access-vj86w\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:53 crc kubenswrapper[4733]: I1204 19:09:53.855098 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99dc23-04af-4f68-b2cc-7f9b5f29798d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:09:54 crc kubenswrapper[4733]: I1204 19:09:54.166008 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f755-account-create-update-sgvzv" event={"ID":"e23e0f7f-8466-4896-9aae-0fd66d607599","Type":"ContainerDied","Data":"d1a54f6ec2bbfa749ed314db099976d1b4d7fec4777af230829c61b22d37b64d"} Dec 04 19:09:54 crc kubenswrapper[4733]: I1204 19:09:54.166280 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1a54f6ec2bbfa749ed314db099976d1b4d7fec4777af230829c61b22d37b64d" Dec 04 19:09:54 crc kubenswrapper[4733]: I1204 19:09:54.166059 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f755-account-create-update-sgvzv" Dec 04 19:09:54 crc kubenswrapper[4733]: I1204 19:09:54.167474 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bsb7q" event={"ID":"4f99dc23-04af-4f68-b2cc-7f9b5f29798d","Type":"ContainerDied","Data":"63d46af368f1ddc31287cbe2422ee53b0806b4af0d5dd3f90f1352388a5790f6"} Dec 04 19:09:54 crc kubenswrapper[4733]: I1204 19:09:54.167534 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63d46af368f1ddc31287cbe2422ee53b0806b4af0d5dd3f90f1352388a5790f6" Dec 04 19:09:54 crc kubenswrapper[4733]: I1204 19:09:54.167593 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bsb7q" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.791766 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-sxc9r"] Dec 04 19:09:55 crc kubenswrapper[4733]: E1204 19:09:55.792643 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23e0f7f-8466-4896-9aae-0fd66d607599" containerName="mariadb-account-create-update" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.792660 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23e0f7f-8466-4896-9aae-0fd66d607599" containerName="mariadb-account-create-update" Dec 04 19:09:55 crc kubenswrapper[4733]: E1204 19:09:55.792675 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f99dc23-04af-4f68-b2cc-7f9b5f29798d" containerName="mariadb-database-create" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.792682 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f99dc23-04af-4f68-b2cc-7f9b5f29798d" containerName="mariadb-database-create" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.792971 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23e0f7f-8466-4896-9aae-0fd66d607599" containerName="mariadb-account-create-update" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.793001 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f99dc23-04af-4f68-b2cc-7f9b5f29798d" containerName="mariadb-database-create" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.793739 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.795820 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.796061 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.796147 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.796239 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxt8c" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.804485 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-sxc9r"] Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.894232 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-combined-ca-bundle\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.894343 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-config-data\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.894382 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2vsj\" (UniqueName: \"kubernetes.io/projected/4f77211d-e7c6-4a42-afd5-663e4a4170d2-kube-api-access-b2vsj\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.995701 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-config-data\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.995764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2vsj\" (UniqueName: \"kubernetes.io/projected/4f77211d-e7c6-4a42-afd5-663e4a4170d2-kube-api-access-b2vsj\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:55 crc kubenswrapper[4733]: I1204 19:09:55.995848 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-combined-ca-bundle\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:56 crc kubenswrapper[4733]: I1204 19:09:56.002123 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-combined-ca-bundle\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:56 crc kubenswrapper[4733]: I1204 19:09:56.003709 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-config-data\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:56 crc kubenswrapper[4733]: I1204 19:09:56.011862 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2vsj\" (UniqueName: \"kubernetes.io/projected/4f77211d-e7c6-4a42-afd5-663e4a4170d2-kube-api-access-b2vsj\") pod \"keystone-db-sync-sxc9r\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:56 crc kubenswrapper[4733]: I1204 19:09:56.114748 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:09:56 crc kubenswrapper[4733]: I1204 19:09:56.562769 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-sxc9r"] Dec 04 19:09:57 crc kubenswrapper[4733]: I1204 19:09:57.192833 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-sxc9r" event={"ID":"4f77211d-e7c6-4a42-afd5-663e4a4170d2","Type":"ContainerStarted","Data":"cd09e372d6385c95f93a88b440c3650f5bbd3b00f569ac6426f887332705565f"} Dec 04 19:09:57 crc kubenswrapper[4733]: I1204 19:09:57.192885 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-sxc9r" event={"ID":"4f77211d-e7c6-4a42-afd5-663e4a4170d2","Type":"ContainerStarted","Data":"5421f2aa8d33006cc273a74ce686ae05ab4d4a150cb3abb3eb72db465eced578"} Dec 04 19:09:57 crc kubenswrapper[4733]: I1204 19:09:57.219552 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-sxc9r" podStartSLOduration=2.21952829 podStartE2EDuration="2.21952829s" podCreationTimestamp="2025-12-04 19:09:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:09:57.210158547 +0000 UTC m=+5459.165519633" watchObservedRunningTime="2025-12-04 19:09:57.21952829 +0000 UTC m=+5459.174889346" Dec 04 19:09:58 crc kubenswrapper[4733]: E1204 19:09:58.360641 4733 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/ccf74bb63478577e1fbaa14259a0209a5f4b663dd25366c3e1a61f0bf0860f2f/diff" to get inode usage: stat /var/lib/containers/storage/overlay/ccf74bb63478577e1fbaa14259a0209a5f4b663dd25366c3e1a61f0bf0860f2f/diff: no such file or directory, extraDiskErr: Dec 04 19:09:59 crc kubenswrapper[4733]: I1204 19:09:59.216677 4733 generic.go:334] "Generic (PLEG): container finished" podID="4f77211d-e7c6-4a42-afd5-663e4a4170d2" containerID="cd09e372d6385c95f93a88b440c3650f5bbd3b00f569ac6426f887332705565f" exitCode=0 Dec 04 19:09:59 crc kubenswrapper[4733]: I1204 19:09:59.216823 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-sxc9r" event={"ID":"4f77211d-e7c6-4a42-afd5-663e4a4170d2","Type":"ContainerDied","Data":"cd09e372d6385c95f93a88b440c3650f5bbd3b00f569ac6426f887332705565f"} Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.037838 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.596299 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.681139 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2vsj\" (UniqueName: \"kubernetes.io/projected/4f77211d-e7c6-4a42-afd5-663e4a4170d2-kube-api-access-b2vsj\") pod \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.681246 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-combined-ca-bundle\") pod \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.681285 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-config-data\") pod \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\" (UID: \"4f77211d-e7c6-4a42-afd5-663e4a4170d2\") " Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.686505 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f77211d-e7c6-4a42-afd5-663e4a4170d2-kube-api-access-b2vsj" (OuterVolumeSpecName: "kube-api-access-b2vsj") pod "4f77211d-e7c6-4a42-afd5-663e4a4170d2" (UID: "4f77211d-e7c6-4a42-afd5-663e4a4170d2"). InnerVolumeSpecName "kube-api-access-b2vsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.740087 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f77211d-e7c6-4a42-afd5-663e4a4170d2" (UID: "4f77211d-e7c6-4a42-afd5-663e4a4170d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.741739 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-config-data" (OuterVolumeSpecName: "config-data") pod "4f77211d-e7c6-4a42-afd5-663e4a4170d2" (UID: "4f77211d-e7c6-4a42-afd5-663e4a4170d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.782966 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2vsj\" (UniqueName: \"kubernetes.io/projected/4f77211d-e7c6-4a42-afd5-663e4a4170d2-kube-api-access-b2vsj\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.782999 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:00 crc kubenswrapper[4733]: I1204 19:10:00.783014 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f77211d-e7c6-4a42-afd5-663e4a4170d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.237598 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-sxc9r" event={"ID":"4f77211d-e7c6-4a42-afd5-663e4a4170d2","Type":"ContainerDied","Data":"5421f2aa8d33006cc273a74ce686ae05ab4d4a150cb3abb3eb72db465eced578"} Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.237642 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5421f2aa8d33006cc273a74ce686ae05ab4d4a150cb3abb3eb72db465eced578" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.237640 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-sxc9r" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.487434 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6f5d79c7-b7qkb"] Dec 04 19:10:01 crc kubenswrapper[4733]: E1204 19:10:01.487890 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f77211d-e7c6-4a42-afd5-663e4a4170d2" containerName="keystone-db-sync" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.487910 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f77211d-e7c6-4a42-afd5-663e4a4170d2" containerName="keystone-db-sync" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.488144 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f77211d-e7c6-4a42-afd5-663e4a4170d2" containerName="keystone-db-sync" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.489257 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.500853 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6f5d79c7-b7qkb"] Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.548402 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7qspx"] Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.549996 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.551754 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.554149 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.554519 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.554158 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxt8c" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.554277 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.556429 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7qspx"] Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.605614 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-dns-svc\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.605668 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.605706 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7zth\" (UniqueName: \"kubernetes.io/projected/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-kube-api-access-k7zth\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.605776 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.605859 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-config\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707566 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7zth\" (UniqueName: \"kubernetes.io/projected/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-kube-api-access-k7zth\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707624 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-combined-ca-bundle\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707665 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707689 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-config\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707711 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-scripts\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707740 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-credential-keys\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707767 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x45r\" (UniqueName: \"kubernetes.io/projected/564a44a0-f895-4c4d-87e7-5a409c91baca-kube-api-access-8x45r\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707828 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-dns-svc\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707852 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-config-data\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707878 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.707894 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-fernet-keys\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.708497 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.708563 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-config\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.708581 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.708609 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-dns-svc\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.725464 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7zth\" (UniqueName: \"kubernetes.io/projected/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-kube-api-access-k7zth\") pod \"dnsmasq-dns-6f6f5d79c7-b7qkb\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.809474 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x45r\" (UniqueName: \"kubernetes.io/projected/564a44a0-f895-4c4d-87e7-5a409c91baca-kube-api-access-8x45r\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.809838 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-config-data\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.809868 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-fernet-keys\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.809904 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-combined-ca-bundle\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.809952 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-scripts\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.809979 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-credential-keys\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.813923 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-scripts\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.813962 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-combined-ca-bundle\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.814899 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-fernet-keys\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.815589 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-config-data\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.821277 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-credential-keys\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.827042 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.827169 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x45r\" (UniqueName: \"kubernetes.io/projected/564a44a0-f895-4c4d-87e7-5a409c91baca-kube-api-access-8x45r\") pod \"keystone-bootstrap-7qspx\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:01 crc kubenswrapper[4733]: I1204 19:10:01.879590 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:02 crc kubenswrapper[4733]: I1204 19:10:02.291295 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6f5d79c7-b7qkb"] Dec 04 19:10:02 crc kubenswrapper[4733]: W1204 19:10:02.299077 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8cd68b2_cfcb_418d_94c1_588ee8bf6404.slice/crio-2dbfc74dd414d066dc900ac5c06a86b57db54df0cb73c1f7f4aa38e3734c8bad WatchSource:0}: Error finding container 2dbfc74dd414d066dc900ac5c06a86b57db54df0cb73c1f7f4aa38e3734c8bad: Status 404 returned error can't find the container with id 2dbfc74dd414d066dc900ac5c06a86b57db54df0cb73c1f7f4aa38e3734c8bad Dec 04 19:10:02 crc kubenswrapper[4733]: W1204 19:10:02.393600 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod564a44a0_f895_4c4d_87e7_5a409c91baca.slice/crio-a6b3384013e8f40cced00ca9f2b503f8c966b53f782b92794bd6abd3c6d1fe25 WatchSource:0}: Error finding container a6b3384013e8f40cced00ca9f2b503f8c966b53f782b92794bd6abd3c6d1fe25: Status 404 returned error can't find the container with id a6b3384013e8f40cced00ca9f2b503f8c966b53f782b92794bd6abd3c6d1fe25 Dec 04 19:10:02 crc kubenswrapper[4733]: I1204 19:10:02.399490 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7qspx"] Dec 04 19:10:03 crc kubenswrapper[4733]: I1204 19:10:03.276116 4733 generic.go:334] "Generic (PLEG): container finished" podID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" containerID="18d01ddc86637621e9b7db8de514c04f7bac1a444ab855a2498cb7bb5d6ec169" exitCode=0 Dec 04 19:10:03 crc kubenswrapper[4733]: I1204 19:10:03.276860 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" event={"ID":"b8cd68b2-cfcb-418d-94c1-588ee8bf6404","Type":"ContainerDied","Data":"18d01ddc86637621e9b7db8de514c04f7bac1a444ab855a2498cb7bb5d6ec169"} Dec 04 19:10:03 crc kubenswrapper[4733]: I1204 19:10:03.277623 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" event={"ID":"b8cd68b2-cfcb-418d-94c1-588ee8bf6404","Type":"ContainerStarted","Data":"2dbfc74dd414d066dc900ac5c06a86b57db54df0cb73c1f7f4aa38e3734c8bad"} Dec 04 19:10:03 crc kubenswrapper[4733]: I1204 19:10:03.282681 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7qspx" event={"ID":"564a44a0-f895-4c4d-87e7-5a409c91baca","Type":"ContainerStarted","Data":"b6631a3850f93db5a752c124cb517422fe9d4abd89bb55b8ddbaced614693bfc"} Dec 04 19:10:03 crc kubenswrapper[4733]: I1204 19:10:03.282733 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7qspx" event={"ID":"564a44a0-f895-4c4d-87e7-5a409c91baca","Type":"ContainerStarted","Data":"a6b3384013e8f40cced00ca9f2b503f8c966b53f782b92794bd6abd3c6d1fe25"} Dec 04 19:10:03 crc kubenswrapper[4733]: I1204 19:10:03.326073 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7qspx" podStartSLOduration=2.326052988 podStartE2EDuration="2.326052988s" podCreationTimestamp="2025-12-04 19:10:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:10:03.323093288 +0000 UTC m=+5465.278454334" watchObservedRunningTime="2025-12-04 19:10:03.326052988 +0000 UTC m=+5465.281414034" Dec 04 19:10:04 crc kubenswrapper[4733]: I1204 19:10:04.295480 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" event={"ID":"b8cd68b2-cfcb-418d-94c1-588ee8bf6404","Type":"ContainerStarted","Data":"b318c7d579d40af5b01c645d49ad63771afb37831a1ddd4763645127cfa6d4a0"} Dec 04 19:10:04 crc kubenswrapper[4733]: I1204 19:10:04.296002 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:04 crc kubenswrapper[4733]: I1204 19:10:04.328547 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" podStartSLOduration=3.328528972 podStartE2EDuration="3.328528972s" podCreationTimestamp="2025-12-04 19:10:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:10:04.32660993 +0000 UTC m=+5466.281970966" watchObservedRunningTime="2025-12-04 19:10:04.328528972 +0000 UTC m=+5466.283890018" Dec 04 19:10:06 crc kubenswrapper[4733]: I1204 19:10:06.314565 4733 generic.go:334] "Generic (PLEG): container finished" podID="564a44a0-f895-4c4d-87e7-5a409c91baca" containerID="b6631a3850f93db5a752c124cb517422fe9d4abd89bb55b8ddbaced614693bfc" exitCode=0 Dec 04 19:10:06 crc kubenswrapper[4733]: I1204 19:10:06.314677 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7qspx" event={"ID":"564a44a0-f895-4c4d-87e7-5a409c91baca","Type":"ContainerDied","Data":"b6631a3850f93db5a752c124cb517422fe9d4abd89bb55b8ddbaced614693bfc"} Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.685548 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.839303 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-config-data\") pod \"564a44a0-f895-4c4d-87e7-5a409c91baca\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.839378 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-scripts\") pod \"564a44a0-f895-4c4d-87e7-5a409c91baca\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.839411 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-credential-keys\") pod \"564a44a0-f895-4c4d-87e7-5a409c91baca\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.839499 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-combined-ca-bundle\") pod \"564a44a0-f895-4c4d-87e7-5a409c91baca\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.839571 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x45r\" (UniqueName: \"kubernetes.io/projected/564a44a0-f895-4c4d-87e7-5a409c91baca-kube-api-access-8x45r\") pod \"564a44a0-f895-4c4d-87e7-5a409c91baca\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.839773 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-fernet-keys\") pod \"564a44a0-f895-4c4d-87e7-5a409c91baca\" (UID: \"564a44a0-f895-4c4d-87e7-5a409c91baca\") " Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.844659 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/564a44a0-f895-4c4d-87e7-5a409c91baca-kube-api-access-8x45r" (OuterVolumeSpecName: "kube-api-access-8x45r") pod "564a44a0-f895-4c4d-87e7-5a409c91baca" (UID: "564a44a0-f895-4c4d-87e7-5a409c91baca"). InnerVolumeSpecName "kube-api-access-8x45r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.845832 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "564a44a0-f895-4c4d-87e7-5a409c91baca" (UID: "564a44a0-f895-4c4d-87e7-5a409c91baca"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.845918 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "564a44a0-f895-4c4d-87e7-5a409c91baca" (UID: "564a44a0-f895-4c4d-87e7-5a409c91baca"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.850911 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-scripts" (OuterVolumeSpecName: "scripts") pod "564a44a0-f895-4c4d-87e7-5a409c91baca" (UID: "564a44a0-f895-4c4d-87e7-5a409c91baca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.875038 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-config-data" (OuterVolumeSpecName: "config-data") pod "564a44a0-f895-4c4d-87e7-5a409c91baca" (UID: "564a44a0-f895-4c4d-87e7-5a409c91baca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.883166 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "564a44a0-f895-4c4d-87e7-5a409c91baca" (UID: "564a44a0-f895-4c4d-87e7-5a409c91baca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.941600 4733 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.941894 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.942035 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x45r\" (UniqueName: \"kubernetes.io/projected/564a44a0-f895-4c4d-87e7-5a409c91baca-kube-api-access-8x45r\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.942158 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.942272 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:07 crc kubenswrapper[4733]: I1204 19:10:07.942417 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/564a44a0-f895-4c4d-87e7-5a409c91baca-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.345336 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7qspx" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.348011 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7qspx" event={"ID":"564a44a0-f895-4c4d-87e7-5a409c91baca","Type":"ContainerDied","Data":"a6b3384013e8f40cced00ca9f2b503f8c966b53f782b92794bd6abd3c6d1fe25"} Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.352110 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6b3384013e8f40cced00ca9f2b503f8c966b53f782b92794bd6abd3c6d1fe25" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.433917 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7qspx"] Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.443988 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7qspx"] Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.528470 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-m5tgq"] Dec 04 19:10:08 crc kubenswrapper[4733]: E1204 19:10:08.528937 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="564a44a0-f895-4c4d-87e7-5a409c91baca" containerName="keystone-bootstrap" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.528963 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="564a44a0-f895-4c4d-87e7-5a409c91baca" containerName="keystone-bootstrap" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.529198 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="564a44a0-f895-4c4d-87e7-5a409c91baca" containerName="keystone-bootstrap" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.529871 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.538041 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m5tgq"] Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.567531 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-scripts\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.567609 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgqrp\" (UniqueName: \"kubernetes.io/projected/09d2f0ad-ae4c-4307-97d8-35ac282147bc-kube-api-access-pgqrp\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.567658 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-credential-keys\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.567692 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-combined-ca-bundle\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.567722 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-fernet-keys\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.567785 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-config-data\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.570532 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.570636 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.570706 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.570756 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxt8c" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.571005 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.669142 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgqrp\" (UniqueName: \"kubernetes.io/projected/09d2f0ad-ae4c-4307-97d8-35ac282147bc-kube-api-access-pgqrp\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.669210 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-credential-keys\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.669249 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-combined-ca-bundle\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.670011 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-fernet-keys\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.670169 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-config-data\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.670339 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-scripts\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.674380 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-credential-keys\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.674550 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-config-data\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.676315 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-combined-ca-bundle\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.676463 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-fernet-keys\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.676835 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-scripts\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.693006 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgqrp\" (UniqueName: \"kubernetes.io/projected/09d2f0ad-ae4c-4307-97d8-35ac282147bc-kube-api-access-pgqrp\") pod \"keystone-bootstrap-m5tgq\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:08 crc kubenswrapper[4733]: I1204 19:10:08.884463 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:09 crc kubenswrapper[4733]: I1204 19:10:09.336442 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m5tgq"] Dec 04 19:10:10 crc kubenswrapper[4733]: I1204 19:10:10.353521 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="564a44a0-f895-4c4d-87e7-5a409c91baca" path="/var/lib/kubelet/pods/564a44a0-f895-4c4d-87e7-5a409c91baca/volumes" Dec 04 19:10:10 crc kubenswrapper[4733]: I1204 19:10:10.362440 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m5tgq" event={"ID":"09d2f0ad-ae4c-4307-97d8-35ac282147bc","Type":"ContainerStarted","Data":"a1a33d66584ca51460258f595aab008766474cb28995466693b8b8de9ba4a26c"} Dec 04 19:10:10 crc kubenswrapper[4733]: I1204 19:10:10.362482 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m5tgq" event={"ID":"09d2f0ad-ae4c-4307-97d8-35ac282147bc","Type":"ContainerStarted","Data":"c89ac3c3db01441883fb849d892163375ef31ec8959925cbce04dfe224a7639a"} Dec 04 19:10:10 crc kubenswrapper[4733]: I1204 19:10:10.385441 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-m5tgq" podStartSLOduration=2.385420027 podStartE2EDuration="2.385420027s" podCreationTimestamp="2025-12-04 19:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:10:10.38033952 +0000 UTC m=+5472.335700606" watchObservedRunningTime="2025-12-04 19:10:10.385420027 +0000 UTC m=+5472.340781083" Dec 04 19:10:11 crc kubenswrapper[4733]: I1204 19:10:11.828562 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:10:11 crc kubenswrapper[4733]: I1204 19:10:11.894664 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66775b999f-sdbwg"] Dec 04 19:10:11 crc kubenswrapper[4733]: I1204 19:10:11.894962 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" podUID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" containerName="dnsmasq-dns" containerID="cri-o://aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87" gracePeriod=10 Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.328304 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.381618 4733 generic.go:334] "Generic (PLEG): container finished" podID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" containerID="aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87" exitCode=0 Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.381706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" event={"ID":"4ff0d2bd-c12a-4ebd-8649-3954979678b3","Type":"ContainerDied","Data":"aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87"} Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.381742 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" event={"ID":"4ff0d2bd-c12a-4ebd-8649-3954979678b3","Type":"ContainerDied","Data":"eb822f5d5385ddace766897adf8751abac5ccb47cec3026efbdcd96872069dc0"} Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.381769 4733 scope.go:117] "RemoveContainer" containerID="aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.381975 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66775b999f-sdbwg" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.384249 4733 generic.go:334] "Generic (PLEG): container finished" podID="09d2f0ad-ae4c-4307-97d8-35ac282147bc" containerID="a1a33d66584ca51460258f595aab008766474cb28995466693b8b8de9ba4a26c" exitCode=0 Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.384295 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m5tgq" event={"ID":"09d2f0ad-ae4c-4307-97d8-35ac282147bc","Type":"ContainerDied","Data":"a1a33d66584ca51460258f595aab008766474cb28995466693b8b8de9ba4a26c"} Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.411902 4733 scope.go:117] "RemoveContainer" containerID="b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.433968 4733 scope.go:117] "RemoveContainer" containerID="aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87" Dec 04 19:10:12 crc kubenswrapper[4733]: E1204 19:10:12.434488 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87\": container with ID starting with aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87 not found: ID does not exist" containerID="aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.434530 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87"} err="failed to get container status \"aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87\": rpc error: code = NotFound desc = could not find container \"aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87\": container with ID starting with aa1ca3db7357e4094b6f5024332e9b210441253dd7ed52b7a86fe3d960aaac87 not found: ID does not exist" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.434556 4733 scope.go:117] "RemoveContainer" containerID="b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656" Dec 04 19:10:12 crc kubenswrapper[4733]: E1204 19:10:12.434946 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656\": container with ID starting with b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656 not found: ID does not exist" containerID="b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.434978 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656"} err="failed to get container status \"b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656\": rpc error: code = NotFound desc = could not find container \"b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656\": container with ID starting with b78e6c9504e08e798ab0a922002c0927fde8dc830f3568527a90e7d0a26cc656 not found: ID does not exist" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.485284 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj5lt\" (UniqueName: \"kubernetes.io/projected/4ff0d2bd-c12a-4ebd-8649-3954979678b3-kube-api-access-zj5lt\") pod \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.485336 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-dns-svc\") pod \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.485473 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-nb\") pod \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.485519 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-config\") pod \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.485554 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-sb\") pod \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\" (UID: \"4ff0d2bd-c12a-4ebd-8649-3954979678b3\") " Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.497359 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff0d2bd-c12a-4ebd-8649-3954979678b3-kube-api-access-zj5lt" (OuterVolumeSpecName: "kube-api-access-zj5lt") pod "4ff0d2bd-c12a-4ebd-8649-3954979678b3" (UID: "4ff0d2bd-c12a-4ebd-8649-3954979678b3"). InnerVolumeSpecName "kube-api-access-zj5lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.532368 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-config" (OuterVolumeSpecName: "config") pod "4ff0d2bd-c12a-4ebd-8649-3954979678b3" (UID: "4ff0d2bd-c12a-4ebd-8649-3954979678b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.539939 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ff0d2bd-c12a-4ebd-8649-3954979678b3" (UID: "4ff0d2bd-c12a-4ebd-8649-3954979678b3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.541137 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ff0d2bd-c12a-4ebd-8649-3954979678b3" (UID: "4ff0d2bd-c12a-4ebd-8649-3954979678b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.543232 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4ff0d2bd-c12a-4ebd-8649-3954979678b3" (UID: "4ff0d2bd-c12a-4ebd-8649-3954979678b3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.588370 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.588445 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.588461 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.588476 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj5lt\" (UniqueName: \"kubernetes.io/projected/4ff0d2bd-c12a-4ebd-8649-3954979678b3-kube-api-access-zj5lt\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.588490 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ff0d2bd-c12a-4ebd-8649-3954979678b3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.719512 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66775b999f-sdbwg"] Dec 04 19:10:12 crc kubenswrapper[4733]: I1204 19:10:12.731508 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66775b999f-sdbwg"] Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.816857 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.925423 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgqrp\" (UniqueName: \"kubernetes.io/projected/09d2f0ad-ae4c-4307-97d8-35ac282147bc-kube-api-access-pgqrp\") pod \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.925485 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-fernet-keys\") pod \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.925535 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-credential-keys\") pod \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.925572 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-config-data\") pod \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.925613 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-scripts\") pod \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.925655 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-combined-ca-bundle\") pod \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\" (UID: \"09d2f0ad-ae4c-4307-97d8-35ac282147bc\") " Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.931399 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-scripts" (OuterVolumeSpecName: "scripts") pod "09d2f0ad-ae4c-4307-97d8-35ac282147bc" (UID: "09d2f0ad-ae4c-4307-97d8-35ac282147bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.931571 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "09d2f0ad-ae4c-4307-97d8-35ac282147bc" (UID: "09d2f0ad-ae4c-4307-97d8-35ac282147bc"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.931922 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09d2f0ad-ae4c-4307-97d8-35ac282147bc-kube-api-access-pgqrp" (OuterVolumeSpecName: "kube-api-access-pgqrp") pod "09d2f0ad-ae4c-4307-97d8-35ac282147bc" (UID: "09d2f0ad-ae4c-4307-97d8-35ac282147bc"). InnerVolumeSpecName "kube-api-access-pgqrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.942724 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "09d2f0ad-ae4c-4307-97d8-35ac282147bc" (UID: "09d2f0ad-ae4c-4307-97d8-35ac282147bc"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.949461 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09d2f0ad-ae4c-4307-97d8-35ac282147bc" (UID: "09d2f0ad-ae4c-4307-97d8-35ac282147bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:13 crc kubenswrapper[4733]: I1204 19:10:13.959318 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-config-data" (OuterVolumeSpecName: "config-data") pod "09d2f0ad-ae4c-4307-97d8-35ac282147bc" (UID: "09d2f0ad-ae4c-4307-97d8-35ac282147bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.027587 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgqrp\" (UniqueName: \"kubernetes.io/projected/09d2f0ad-ae4c-4307-97d8-35ac282147bc-kube-api-access-pgqrp\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.027646 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.027660 4733 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.027670 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.027682 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.027694 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d2f0ad-ae4c-4307-97d8-35ac282147bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.347018 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" path="/var/lib/kubelet/pods/4ff0d2bd-c12a-4ebd-8649-3954979678b3/volumes" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.416937 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m5tgq" event={"ID":"09d2f0ad-ae4c-4307-97d8-35ac282147bc","Type":"ContainerDied","Data":"c89ac3c3db01441883fb849d892163375ef31ec8959925cbce04dfe224a7639a"} Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.416997 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c89ac3c3db01441883fb849d892163375ef31ec8959925cbce04dfe224a7639a" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.416966 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m5tgq" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.499668 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5f9c8c86c8-h296q"] Dec 04 19:10:14 crc kubenswrapper[4733]: E1204 19:10:14.500124 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" containerName="dnsmasq-dns" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.500148 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" containerName="dnsmasq-dns" Dec 04 19:10:14 crc kubenswrapper[4733]: E1204 19:10:14.500168 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09d2f0ad-ae4c-4307-97d8-35ac282147bc" containerName="keystone-bootstrap" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.500179 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="09d2f0ad-ae4c-4307-97d8-35ac282147bc" containerName="keystone-bootstrap" Dec 04 19:10:14 crc kubenswrapper[4733]: E1204 19:10:14.500194 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" containerName="init" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.500203 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" containerName="init" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.500399 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="09d2f0ad-ae4c-4307-97d8-35ac282147bc" containerName="keystone-bootstrap" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.500427 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ff0d2bd-c12a-4ebd-8649-3954979678b3" containerName="dnsmasq-dns" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.501146 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.504188 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.504378 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.504490 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.506832 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxt8c" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.561413 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5f9c8c86c8-h296q"] Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.640901 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-fernet-keys\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.640940 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-scripts\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.640984 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-combined-ca-bundle\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.641222 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-credential-keys\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.641286 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-config-data\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.641501 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4rt8\" (UniqueName: \"kubernetes.io/projected/923bc92b-30a9-48c8-9228-46f7e257db9a-kube-api-access-p4rt8\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.742695 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-credential-keys\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.742747 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-config-data\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.742809 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4rt8\" (UniqueName: \"kubernetes.io/projected/923bc92b-30a9-48c8-9228-46f7e257db9a-kube-api-access-p4rt8\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.742839 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-fernet-keys\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.742854 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-scripts\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.742907 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-combined-ca-bundle\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.746489 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-scripts\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.746723 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-credential-keys\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.746843 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-fernet-keys\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.746844 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-config-data\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.748253 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923bc92b-30a9-48c8-9228-46f7e257db9a-combined-ca-bundle\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.768086 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4rt8\" (UniqueName: \"kubernetes.io/projected/923bc92b-30a9-48c8-9228-46f7e257db9a-kube-api-access-p4rt8\") pod \"keystone-5f9c8c86c8-h296q\" (UID: \"923bc92b-30a9-48c8-9228-46f7e257db9a\") " pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:14 crc kubenswrapper[4733]: I1204 19:10:14.826962 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:15 crc kubenswrapper[4733]: I1204 19:10:15.303422 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5f9c8c86c8-h296q"] Dec 04 19:10:15 crc kubenswrapper[4733]: I1204 19:10:15.428593 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f9c8c86c8-h296q" event={"ID":"923bc92b-30a9-48c8-9228-46f7e257db9a","Type":"ContainerStarted","Data":"3de36e97347a0bb879278b4923cbf5ddc8bd6cb105d81abe0cbdb25d3ab398d9"} Dec 04 19:10:16 crc kubenswrapper[4733]: I1204 19:10:16.440950 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f9c8c86c8-h296q" event={"ID":"923bc92b-30a9-48c8-9228-46f7e257db9a","Type":"ContainerStarted","Data":"e8aeedff5e0c369e0bfd03cb2e8382665ca376973a7977bc3219cb59a8a275f7"} Dec 04 19:10:16 crc kubenswrapper[4733]: I1204 19:10:16.441113 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:21 crc kubenswrapper[4733]: I1204 19:10:21.501527 4733 scope.go:117] "RemoveContainer" containerID="ebfbed019bbe9565def43557bc96698a4c8ab45cf1c61fa9badc9e3ae490f7d8" Dec 04 19:10:21 crc kubenswrapper[4733]: I1204 19:10:21.526734 4733 scope.go:117] "RemoveContainer" containerID="8baccd9d860b414cabb3e3a42c4a3eef8ec9f39ca4a5b88fef7a428dc8b71a1e" Dec 04 19:10:46 crc kubenswrapper[4733]: I1204 19:10:46.303592 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5f9c8c86c8-h296q" Dec 04 19:10:46 crc kubenswrapper[4733]: I1204 19:10:46.330727 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5f9c8c86c8-h296q" podStartSLOduration=32.330710466 podStartE2EDuration="32.330710466s" podCreationTimestamp="2025-12-04 19:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:10:16.463185075 +0000 UTC m=+5478.418546121" watchObservedRunningTime="2025-12-04 19:10:46.330710466 +0000 UTC m=+5508.286071502" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.227881 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.229336 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.234600 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.236260 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jltxr" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.238146 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.246654 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.276315 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 04 19:10:50 crc kubenswrapper[4733]: E1204 19:10:50.277005 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-4kvrc openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="5c5ad2a0-ead8-4a03-8433-c5f99ee917b6" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.285702 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.294230 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.295780 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.302218 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.339995 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config-secret\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.340109 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.340312 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kvrc\" (UniqueName: \"kubernetes.io/projected/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-kube-api-access-4kvrc\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.347004 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c5ad2a0-ead8-4a03-8433-c5f99ee917b6" path="/var/lib/kubelet/pods/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6/volumes" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.441603 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.441758 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kvrc\" (UniqueName: \"kubernetes.io/projected/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-kube-api-access-4kvrc\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.441829 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.441902 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config-secret\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.442048 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.442137 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmrs4\" (UniqueName: \"kubernetes.io/projected/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-kube-api-access-pmrs4\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.443182 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: E1204 19:10:50.444210 4733 projected.go:194] Error preparing data for projected volume kube-api-access-4kvrc for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (5c5ad2a0-ead8-4a03-8433-c5f99ee917b6) does not match the UID in record. The object might have been deleted and then recreated Dec 04 19:10:50 crc kubenswrapper[4733]: E1204 19:10:50.444269 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-kube-api-access-4kvrc podName:5c5ad2a0-ead8-4a03-8433-c5f99ee917b6 nodeName:}" failed. No retries permitted until 2025-12-04 19:10:50.944252546 +0000 UTC m=+5512.899613592 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-4kvrc" (UniqueName: "kubernetes.io/projected/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-kube-api-access-4kvrc") pod "openstackclient" (UID: "5c5ad2a0-ead8-4a03-8433-c5f99ee917b6") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (5c5ad2a0-ead8-4a03-8433-c5f99ee917b6) does not match the UID in record. The object might have been deleted and then recreated Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.448088 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config-secret\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.544077 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.544443 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmrs4\" (UniqueName: \"kubernetes.io/projected/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-kube-api-access-pmrs4\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.544472 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.545292 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.547433 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.565525 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmrs4\" (UniqueName: \"kubernetes.io/projected/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-kube-api-access-pmrs4\") pod \"openstackclient\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.622814 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.951767 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kvrc\" (UniqueName: \"kubernetes.io/projected/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-kube-api-access-4kvrc\") pod \"openstackclient\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " pod="openstack/openstackclient" Dec 04 19:10:50 crc kubenswrapper[4733]: I1204 19:10:50.954056 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 19:10:50 crc kubenswrapper[4733]: E1204 19:10:50.954109 4733 projected.go:194] Error preparing data for projected volume kube-api-access-4kvrc for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (5c5ad2a0-ead8-4a03-8433-c5f99ee917b6) does not match the UID in record. The object might have been deleted and then recreated Dec 04 19:10:50 crc kubenswrapper[4733]: E1204 19:10:50.954375 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-kube-api-access-4kvrc podName:5c5ad2a0-ead8-4a03-8433-c5f99ee917b6 nodeName:}" failed. No retries permitted until 2025-12-04 19:10:51.954160473 +0000 UTC m=+5513.909521509 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-4kvrc" (UniqueName: "kubernetes.io/projected/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-kube-api-access-4kvrc") pod "openstackclient" (UID: "5c5ad2a0-ead8-4a03-8433-c5f99ee917b6") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (5c5ad2a0-ead8-4a03-8433-c5f99ee917b6) does not match the UID in record. The object might have been deleted and then recreated Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.162563 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.162563 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5","Type":"ContainerStarted","Data":"28e79a2f25597ff7c04d43e40505344eadc98163b4e687c090db85ca610569c9"} Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.162649 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5","Type":"ContainerStarted","Data":"abc82d31ca0ca39acee0bfbe4717f80d736100bde3964f83f1a82424a90c0bcf"} Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.177303 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.177285336 podStartE2EDuration="1.177285336s" podCreationTimestamp="2025-12-04 19:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:10:51.173820493 +0000 UTC m=+5513.129181569" watchObservedRunningTime="2025-12-04 19:10:51.177285336 +0000 UTC m=+5513.132646382" Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.177676 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.180266 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5c5ad2a0-ead8-4a03-8433-c5f99ee917b6" podUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.358697 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config-secret\") pod \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.358918 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config\") pod \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\" (UID: \"5c5ad2a0-ead8-4a03-8433-c5f99ee917b6\") " Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.359343 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kvrc\" (UniqueName: \"kubernetes.io/projected/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-kube-api-access-4kvrc\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.360039 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5c5ad2a0-ead8-4a03-8433-c5f99ee917b6" (UID: "5c5ad2a0-ead8-4a03-8433-c5f99ee917b6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.362342 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5c5ad2a0-ead8-4a03-8433-c5f99ee917b6" (UID: "5c5ad2a0-ead8-4a03-8433-c5f99ee917b6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.461507 4733 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:51 crc kubenswrapper[4733]: I1204 19:10:51.461546 4733 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 19:10:52 crc kubenswrapper[4733]: I1204 19:10:52.170135 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:10:52 crc kubenswrapper[4733]: I1204 19:10:52.187608 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5c5ad2a0-ead8-4a03-8433-c5f99ee917b6" podUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" Dec 04 19:10:52 crc kubenswrapper[4733]: I1204 19:10:52.347599 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c5ad2a0-ead8-4a03-8433-c5f99ee917b6" path="/var/lib/kubelet/pods/5c5ad2a0-ead8-4a03-8433-c5f99ee917b6/volumes" Dec 04 19:11:15 crc kubenswrapper[4733]: I1204 19:11:15.361650 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:11:15 crc kubenswrapper[4733]: I1204 19:11:15.362501 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:11:45 crc kubenswrapper[4733]: I1204 19:11:45.362704 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:11:45 crc kubenswrapper[4733]: I1204 19:11:45.363444 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:12:15 crc kubenswrapper[4733]: I1204 19:12:15.362275 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:12:15 crc kubenswrapper[4733]: I1204 19:12:15.362925 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:12:15 crc kubenswrapper[4733]: I1204 19:12:15.362981 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:12:15 crc kubenswrapper[4733]: I1204 19:12:15.363628 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef4c43597b7d9b6a5f9198424285408b276a51a3896ec710afa65d9b6bbf8b2b"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:12:15 crc kubenswrapper[4733]: I1204 19:12:15.363701 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://ef4c43597b7d9b6a5f9198424285408b276a51a3896ec710afa65d9b6bbf8b2b" gracePeriod=600 Dec 04 19:12:16 crc kubenswrapper[4733]: I1204 19:12:16.245213 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="ef4c43597b7d9b6a5f9198424285408b276a51a3896ec710afa65d9b6bbf8b2b" exitCode=0 Dec 04 19:12:16 crc kubenswrapper[4733]: I1204 19:12:16.245298 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"ef4c43597b7d9b6a5f9198424285408b276a51a3896ec710afa65d9b6bbf8b2b"} Dec 04 19:12:16 crc kubenswrapper[4733]: I1204 19:12:16.246168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0"} Dec 04 19:12:16 crc kubenswrapper[4733]: I1204 19:12:16.246205 4733 scope.go:117] "RemoveContainer" containerID="65083e912f2f2a0e3501c566953a8c0db3bd81766cfae57062c599f9b4f8dbaf" Dec 04 19:12:21 crc kubenswrapper[4733]: E1204 19:12:21.043118 4733 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.246:41322->38.102.83.246:40229: read tcp 38.102.83.246:41322->38.102.83.246:40229: read: connection reset by peer Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.609505 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-7s2bl"] Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.611849 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.615868 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-983b-account-create-update-gr7xd"] Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.616783 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.618267 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.621555 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-983b-account-create-update-gr7xd"] Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.627376 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7s2bl"] Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.717222 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7104c6-4073-40e7-b906-a659739adbb1-operator-scripts\") pod \"barbican-983b-account-create-update-gr7xd\" (UID: \"9f7104c6-4073-40e7-b906-a659739adbb1\") " pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.717294 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmvgn\" (UniqueName: \"kubernetes.io/projected/40f10d39-9d06-4362-b9f8-d4e05f623cdc-kube-api-access-xmvgn\") pod \"barbican-db-create-7s2bl\" (UID: \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\") " pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.717399 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40f10d39-9d06-4362-b9f8-d4e05f623cdc-operator-scripts\") pod \"barbican-db-create-7s2bl\" (UID: \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\") " pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.717514 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-654wv\" (UniqueName: \"kubernetes.io/projected/9f7104c6-4073-40e7-b906-a659739adbb1-kube-api-access-654wv\") pod \"barbican-983b-account-create-update-gr7xd\" (UID: \"9f7104c6-4073-40e7-b906-a659739adbb1\") " pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.819670 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmvgn\" (UniqueName: \"kubernetes.io/projected/40f10d39-9d06-4362-b9f8-d4e05f623cdc-kube-api-access-xmvgn\") pod \"barbican-db-create-7s2bl\" (UID: \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\") " pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.819759 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40f10d39-9d06-4362-b9f8-d4e05f623cdc-operator-scripts\") pod \"barbican-db-create-7s2bl\" (UID: \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\") " pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.819827 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-654wv\" (UniqueName: \"kubernetes.io/projected/9f7104c6-4073-40e7-b906-a659739adbb1-kube-api-access-654wv\") pod \"barbican-983b-account-create-update-gr7xd\" (UID: \"9f7104c6-4073-40e7-b906-a659739adbb1\") " pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.819963 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7104c6-4073-40e7-b906-a659739adbb1-operator-scripts\") pod \"barbican-983b-account-create-update-gr7xd\" (UID: \"9f7104c6-4073-40e7-b906-a659739adbb1\") " pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.820701 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40f10d39-9d06-4362-b9f8-d4e05f623cdc-operator-scripts\") pod \"barbican-db-create-7s2bl\" (UID: \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\") " pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.820914 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7104c6-4073-40e7-b906-a659739adbb1-operator-scripts\") pod \"barbican-983b-account-create-update-gr7xd\" (UID: \"9f7104c6-4073-40e7-b906-a659739adbb1\") " pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.840842 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmvgn\" (UniqueName: \"kubernetes.io/projected/40f10d39-9d06-4362-b9f8-d4e05f623cdc-kube-api-access-xmvgn\") pod \"barbican-db-create-7s2bl\" (UID: \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\") " pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.841399 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-654wv\" (UniqueName: \"kubernetes.io/projected/9f7104c6-4073-40e7-b906-a659739adbb1-kube-api-access-654wv\") pod \"barbican-983b-account-create-update-gr7xd\" (UID: \"9f7104c6-4073-40e7-b906-a659739adbb1\") " pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.947499 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:26 crc kubenswrapper[4733]: I1204 19:12:26.956767 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:27 crc kubenswrapper[4733]: I1204 19:12:27.394737 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7s2bl"] Dec 04 19:12:27 crc kubenswrapper[4733]: W1204 19:12:27.400588 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40f10d39_9d06_4362_b9f8_d4e05f623cdc.slice/crio-93b2562f336843a79e07c503c591a979c3dd4efc571a4acaf5fe306a3cb75796 WatchSource:0}: Error finding container 93b2562f336843a79e07c503c591a979c3dd4efc571a4acaf5fe306a3cb75796: Status 404 returned error can't find the container with id 93b2562f336843a79e07c503c591a979c3dd4efc571a4acaf5fe306a3cb75796 Dec 04 19:12:27 crc kubenswrapper[4733]: I1204 19:12:27.450127 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-983b-account-create-update-gr7xd"] Dec 04 19:12:28 crc kubenswrapper[4733]: I1204 19:12:28.353664 4733 generic.go:334] "Generic (PLEG): container finished" podID="9f7104c6-4073-40e7-b906-a659739adbb1" containerID="9fb8361d9617c8ab101d44217631084e02369a3904f663990070e5df1e5f69b0" exitCode=0 Dec 04 19:12:28 crc kubenswrapper[4733]: I1204 19:12:28.353729 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-983b-account-create-update-gr7xd" event={"ID":"9f7104c6-4073-40e7-b906-a659739adbb1","Type":"ContainerDied","Data":"9fb8361d9617c8ab101d44217631084e02369a3904f663990070e5df1e5f69b0"} Dec 04 19:12:28 crc kubenswrapper[4733]: I1204 19:12:28.354232 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-983b-account-create-update-gr7xd" event={"ID":"9f7104c6-4073-40e7-b906-a659739adbb1","Type":"ContainerStarted","Data":"d239be3a38c835aa1392b39be57d4ac45068e9ecfe19d463096c9d33add652e8"} Dec 04 19:12:28 crc kubenswrapper[4733]: I1204 19:12:28.359932 4733 generic.go:334] "Generic (PLEG): container finished" podID="40f10d39-9d06-4362-b9f8-d4e05f623cdc" containerID="950263320cd4ce323201835602127727776b8fdec7eb2e49ce2e935611928b1c" exitCode=0 Dec 04 19:12:28 crc kubenswrapper[4733]: I1204 19:12:28.360002 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7s2bl" event={"ID":"40f10d39-9d06-4362-b9f8-d4e05f623cdc","Type":"ContainerDied","Data":"950263320cd4ce323201835602127727776b8fdec7eb2e49ce2e935611928b1c"} Dec 04 19:12:28 crc kubenswrapper[4733]: I1204 19:12:28.360080 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7s2bl" event={"ID":"40f10d39-9d06-4362-b9f8-d4e05f623cdc","Type":"ContainerStarted","Data":"93b2562f336843a79e07c503c591a979c3dd4efc571a4acaf5fe306a3cb75796"} Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.721936 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.728252 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.768931 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7104c6-4073-40e7-b906-a659739adbb1-operator-scripts\") pod \"9f7104c6-4073-40e7-b906-a659739adbb1\" (UID: \"9f7104c6-4073-40e7-b906-a659739adbb1\") " Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.769082 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40f10d39-9d06-4362-b9f8-d4e05f623cdc-operator-scripts\") pod \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\" (UID: \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\") " Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.769132 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-654wv\" (UniqueName: \"kubernetes.io/projected/9f7104c6-4073-40e7-b906-a659739adbb1-kube-api-access-654wv\") pod \"9f7104c6-4073-40e7-b906-a659739adbb1\" (UID: \"9f7104c6-4073-40e7-b906-a659739adbb1\") " Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.769664 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40f10d39-9d06-4362-b9f8-d4e05f623cdc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "40f10d39-9d06-4362-b9f8-d4e05f623cdc" (UID: "40f10d39-9d06-4362-b9f8-d4e05f623cdc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.769820 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f7104c6-4073-40e7-b906-a659739adbb1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9f7104c6-4073-40e7-b906-a659739adbb1" (UID: "9f7104c6-4073-40e7-b906-a659739adbb1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.775541 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f7104c6-4073-40e7-b906-a659739adbb1-kube-api-access-654wv" (OuterVolumeSpecName: "kube-api-access-654wv") pod "9f7104c6-4073-40e7-b906-a659739adbb1" (UID: "9f7104c6-4073-40e7-b906-a659739adbb1"). InnerVolumeSpecName "kube-api-access-654wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.870736 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmvgn\" (UniqueName: \"kubernetes.io/projected/40f10d39-9d06-4362-b9f8-d4e05f623cdc-kube-api-access-xmvgn\") pod \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\" (UID: \"40f10d39-9d06-4362-b9f8-d4e05f623cdc\") " Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.871549 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40f10d39-9d06-4362-b9f8-d4e05f623cdc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.871786 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-654wv\" (UniqueName: \"kubernetes.io/projected/9f7104c6-4073-40e7-b906-a659739adbb1-kube-api-access-654wv\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.871899 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f7104c6-4073-40e7-b906-a659739adbb1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.873409 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40f10d39-9d06-4362-b9f8-d4e05f623cdc-kube-api-access-xmvgn" (OuterVolumeSpecName: "kube-api-access-xmvgn") pod "40f10d39-9d06-4362-b9f8-d4e05f623cdc" (UID: "40f10d39-9d06-4362-b9f8-d4e05f623cdc"). InnerVolumeSpecName "kube-api-access-xmvgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:12:29 crc kubenswrapper[4733]: I1204 19:12:29.973733 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmvgn\" (UniqueName: \"kubernetes.io/projected/40f10d39-9d06-4362-b9f8-d4e05f623cdc-kube-api-access-xmvgn\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:30 crc kubenswrapper[4733]: I1204 19:12:30.376699 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7s2bl" Dec 04 19:12:30 crc kubenswrapper[4733]: I1204 19:12:30.376780 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7s2bl" event={"ID":"40f10d39-9d06-4362-b9f8-d4e05f623cdc","Type":"ContainerDied","Data":"93b2562f336843a79e07c503c591a979c3dd4efc571a4acaf5fe306a3cb75796"} Dec 04 19:12:30 crc kubenswrapper[4733]: I1204 19:12:30.377222 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93b2562f336843a79e07c503c591a979c3dd4efc571a4acaf5fe306a3cb75796" Dec 04 19:12:30 crc kubenswrapper[4733]: I1204 19:12:30.378916 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-983b-account-create-update-gr7xd" event={"ID":"9f7104c6-4073-40e7-b906-a659739adbb1","Type":"ContainerDied","Data":"d239be3a38c835aa1392b39be57d4ac45068e9ecfe19d463096c9d33add652e8"} Dec 04 19:12:30 crc kubenswrapper[4733]: I1204 19:12:30.378949 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-983b-account-create-update-gr7xd" Dec 04 19:12:30 crc kubenswrapper[4733]: I1204 19:12:30.378959 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d239be3a38c835aa1392b39be57d4ac45068e9ecfe19d463096c9d33add652e8" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.878919 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-2pcqw"] Dec 04 19:12:31 crc kubenswrapper[4733]: E1204 19:12:31.879314 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f7104c6-4073-40e7-b906-a659739adbb1" containerName="mariadb-account-create-update" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.879331 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f7104c6-4073-40e7-b906-a659739adbb1" containerName="mariadb-account-create-update" Dec 04 19:12:31 crc kubenswrapper[4733]: E1204 19:12:31.879355 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f10d39-9d06-4362-b9f8-d4e05f623cdc" containerName="mariadb-database-create" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.879363 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f10d39-9d06-4362-b9f8-d4e05f623cdc" containerName="mariadb-database-create" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.879541 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f7104c6-4073-40e7-b906-a659739adbb1" containerName="mariadb-account-create-update" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.879563 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="40f10d39-9d06-4362-b9f8-d4e05f623cdc" containerName="mariadb-database-create" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.880205 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.882698 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.883757 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-shxd2" Dec 04 19:12:31 crc kubenswrapper[4733]: I1204 19:12:31.896987 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2pcqw"] Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.006335 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-combined-ca-bundle\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.006600 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvzw9\" (UniqueName: \"kubernetes.io/projected/57823d82-7ca6-48c9-8dbf-2f8f3360919d-kube-api-access-xvzw9\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.006684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-db-sync-config-data\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.108687 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-combined-ca-bundle\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.109456 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvzw9\" (UniqueName: \"kubernetes.io/projected/57823d82-7ca6-48c9-8dbf-2f8f3360919d-kube-api-access-xvzw9\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.109662 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-db-sync-config-data\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.115532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-combined-ca-bundle\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.125623 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-db-sync-config-data\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.126651 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvzw9\" (UniqueName: \"kubernetes.io/projected/57823d82-7ca6-48c9-8dbf-2f8f3360919d-kube-api-access-xvzw9\") pod \"barbican-db-sync-2pcqw\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.196784 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:32 crc kubenswrapper[4733]: I1204 19:12:32.702167 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2pcqw"] Dec 04 19:12:32 crc kubenswrapper[4733]: W1204 19:12:32.709981 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57823d82_7ca6_48c9_8dbf_2f8f3360919d.slice/crio-b306717b999f5dbcedd1998c5848d66d7333ab92fd6b9b03fd438ffa6b4a2699 WatchSource:0}: Error finding container b306717b999f5dbcedd1998c5848d66d7333ab92fd6b9b03fd438ffa6b4a2699: Status 404 returned error can't find the container with id b306717b999f5dbcedd1998c5848d66d7333ab92fd6b9b03fd438ffa6b4a2699 Dec 04 19:12:33 crc kubenswrapper[4733]: I1204 19:12:33.403053 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2pcqw" event={"ID":"57823d82-7ca6-48c9-8dbf-2f8f3360919d","Type":"ContainerStarted","Data":"084acd82ccfad183284260343c3ddeeea95837054db447011fdf86650f727c66"} Dec 04 19:12:33 crc kubenswrapper[4733]: I1204 19:12:33.403632 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2pcqw" event={"ID":"57823d82-7ca6-48c9-8dbf-2f8f3360919d","Type":"ContainerStarted","Data":"b306717b999f5dbcedd1998c5848d66d7333ab92fd6b9b03fd438ffa6b4a2699"} Dec 04 19:12:33 crc kubenswrapper[4733]: I1204 19:12:33.424355 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-2pcqw" podStartSLOduration=2.424338019 podStartE2EDuration="2.424338019s" podCreationTimestamp="2025-12-04 19:12:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:12:33.422106339 +0000 UTC m=+5615.377467405" watchObservedRunningTime="2025-12-04 19:12:33.424338019 +0000 UTC m=+5615.379699065" Dec 04 19:12:34 crc kubenswrapper[4733]: I1204 19:12:34.415598 4733 generic.go:334] "Generic (PLEG): container finished" podID="57823d82-7ca6-48c9-8dbf-2f8f3360919d" containerID="084acd82ccfad183284260343c3ddeeea95837054db447011fdf86650f727c66" exitCode=0 Dec 04 19:12:34 crc kubenswrapper[4733]: I1204 19:12:34.415859 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2pcqw" event={"ID":"57823d82-7ca6-48c9-8dbf-2f8f3360919d","Type":"ContainerDied","Data":"084acd82ccfad183284260343c3ddeeea95837054db447011fdf86650f727c66"} Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.805302 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.868509 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-db-sync-config-data\") pod \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.869165 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-combined-ca-bundle\") pod \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.869365 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvzw9\" (UniqueName: \"kubernetes.io/projected/57823d82-7ca6-48c9-8dbf-2f8f3360919d-kube-api-access-xvzw9\") pod \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\" (UID: \"57823d82-7ca6-48c9-8dbf-2f8f3360919d\") " Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.876550 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "57823d82-7ca6-48c9-8dbf-2f8f3360919d" (UID: "57823d82-7ca6-48c9-8dbf-2f8f3360919d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.879119 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57823d82-7ca6-48c9-8dbf-2f8f3360919d-kube-api-access-xvzw9" (OuterVolumeSpecName: "kube-api-access-xvzw9") pod "57823d82-7ca6-48c9-8dbf-2f8f3360919d" (UID: "57823d82-7ca6-48c9-8dbf-2f8f3360919d"). InnerVolumeSpecName "kube-api-access-xvzw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.917319 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57823d82-7ca6-48c9-8dbf-2f8f3360919d" (UID: "57823d82-7ca6-48c9-8dbf-2f8f3360919d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.971916 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.971956 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvzw9\" (UniqueName: \"kubernetes.io/projected/57823d82-7ca6-48c9-8dbf-2f8f3360919d-kube-api-access-xvzw9\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:35 crc kubenswrapper[4733]: I1204 19:12:35.971970 4733 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57823d82-7ca6-48c9-8dbf-2f8f3360919d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.445057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2pcqw" event={"ID":"57823d82-7ca6-48c9-8dbf-2f8f3360919d","Type":"ContainerDied","Data":"b306717b999f5dbcedd1998c5848d66d7333ab92fd6b9b03fd438ffa6b4a2699"} Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.445120 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b306717b999f5dbcedd1998c5848d66d7333ab92fd6b9b03fd438ffa6b4a2699" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.445135 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2pcqw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.690316 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5bc7fbd458-mz9t9"] Dec 04 19:12:36 crc kubenswrapper[4733]: E1204 19:12:36.690748 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57823d82-7ca6-48c9-8dbf-2f8f3360919d" containerName="barbican-db-sync" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.690772 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="57823d82-7ca6-48c9-8dbf-2f8f3360919d" containerName="barbican-db-sync" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.690992 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="57823d82-7ca6-48c9-8dbf-2f8f3360919d" containerName="barbican-db-sync" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.698354 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.699859 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.700899 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-shxd2" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.701241 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.705130 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-68d4f54797-jj274"] Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.710500 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.715468 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.721987 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5bc7fbd458-mz9t9"] Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.730432 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68d4f54797-jj274"] Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786068 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-config-data\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786168 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-config-data-custom\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786213 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb96k\" (UniqueName: \"kubernetes.io/projected/a3871247-7c7c-4ccc-8c58-673cb82b7e97-kube-api-access-gb96k\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786265 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-logs\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786308 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-config-data-custom\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786381 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-combined-ca-bundle\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786434 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xthr\" (UniqueName: \"kubernetes.io/projected/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-kube-api-access-6xthr\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-config-data\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786508 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3871247-7c7c-4ccc-8c58-673cb82b7e97-logs\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.786534 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-combined-ca-bundle\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.815194 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dfdb7c58c-vpkhw"] Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.818469 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.833569 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dfdb7c58c-vpkhw"] Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.876636 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-55cd5f5578-jr2bf"] Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.878651 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.880906 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887721 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xthr\" (UniqueName: \"kubernetes.io/projected/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-kube-api-access-6xthr\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887761 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-dns-svc\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887784 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-config\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887817 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887838 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-config-data\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887854 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3871247-7c7c-4ccc-8c58-673cb82b7e97-logs\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887870 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-combined-ca-bundle\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887904 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-config-data\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887935 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-config-data-custom\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887953 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887974 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb96k\" (UniqueName: \"kubernetes.io/projected/a3871247-7c7c-4ccc-8c58-673cb82b7e97-kube-api-access-gb96k\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.887999 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-logs\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.888024 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-config-data-custom\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.888065 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-combined-ca-bundle\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.888083 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms8cq\" (UniqueName: \"kubernetes.io/projected/3f65e198-57fe-468d-8779-4c25bad24c54-kube-api-access-ms8cq\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.889106 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3871247-7c7c-4ccc-8c58-673cb82b7e97-logs\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.889563 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-logs\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.894457 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-config-data-custom\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.894580 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-config-data\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.895153 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-combined-ca-bundle\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.899558 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3871247-7c7c-4ccc-8c58-673cb82b7e97-combined-ca-bundle\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.911309 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xthr\" (UniqueName: \"kubernetes.io/projected/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-kube-api-access-6xthr\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.911834 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-config-data-custom\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.914373 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66a0bca5-dfe9-40b9-9d30-da5c46f887a2-config-data\") pod \"barbican-worker-68d4f54797-jj274\" (UID: \"66a0bca5-dfe9-40b9-9d30-da5c46f887a2\") " pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.914467 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55cd5f5578-jr2bf"] Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.915965 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb96k\" (UniqueName: \"kubernetes.io/projected/a3871247-7c7c-4ccc-8c58-673cb82b7e97-kube-api-access-gb96k\") pod \"barbican-keystone-listener-5bc7fbd458-mz9t9\" (UID: \"a3871247-7c7c-4ccc-8c58-673cb82b7e97\") " pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.989494 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-config-data\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.989561 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-logs\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.989590 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.989766 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m79t7\" (UniqueName: \"kubernetes.io/projected/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-kube-api-access-m79t7\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.989938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-config-data-custom\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.990025 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms8cq\" (UniqueName: \"kubernetes.io/projected/3f65e198-57fe-468d-8779-4c25bad24c54-kube-api-access-ms8cq\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.990121 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-dns-svc\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.990162 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-config\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.990185 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.990265 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-combined-ca-bundle\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.990499 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.991157 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-dns-svc\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.991553 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-config\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:36 crc kubenswrapper[4733]: I1204 19:12:36.991577 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.007772 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms8cq\" (UniqueName: \"kubernetes.io/projected/3f65e198-57fe-468d-8779-4c25bad24c54-kube-api-access-ms8cq\") pod \"dnsmasq-dns-5dfdb7c58c-vpkhw\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.019444 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.037262 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68d4f54797-jj274" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.092766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-combined-ca-bundle\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.092915 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-config-data\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.092980 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-logs\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.093084 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m79t7\" (UniqueName: \"kubernetes.io/projected/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-kube-api-access-m79t7\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.093118 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-config-data-custom\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.094821 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-logs\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.096017 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-combined-ca-bundle\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.096165 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-config-data-custom\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.099858 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-config-data\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.113741 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m79t7\" (UniqueName: \"kubernetes.io/projected/ae040e65-3459-4cdc-a35b-e2a47f6bcce3-kube-api-access-m79t7\") pod \"barbican-api-55cd5f5578-jr2bf\" (UID: \"ae040e65-3459-4cdc-a35b-e2a47f6bcce3\") " pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.141608 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.275324 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.566046 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5bc7fbd458-mz9t9"] Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.595077 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68d4f54797-jj274"] Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.695226 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55cd5f5578-jr2bf"] Dec 04 19:12:37 crc kubenswrapper[4733]: I1204 19:12:37.701590 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dfdb7c58c-vpkhw"] Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.459520 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68d4f54797-jj274" event={"ID":"66a0bca5-dfe9-40b9-9d30-da5c46f887a2","Type":"ContainerStarted","Data":"5bae05489c65078ea3e97304c0da31b501d6323dd08f713ed093e3d05064bcba"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.459818 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68d4f54797-jj274" event={"ID":"66a0bca5-dfe9-40b9-9d30-da5c46f887a2","Type":"ContainerStarted","Data":"538e0bb513686906ca010c030c01e91b3afb1d7e1fe0face302eb285a1e6bce6"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.459833 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68d4f54797-jj274" event={"ID":"66a0bca5-dfe9-40b9-9d30-da5c46f887a2","Type":"ContainerStarted","Data":"4e7d4a4b8608191acf00d5fb823d302922dc1bb034b017353af8abd336b5d7d8"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.466944 4733 generic.go:334] "Generic (PLEG): container finished" podID="3f65e198-57fe-468d-8779-4c25bad24c54" containerID="0e8c013c0cd2969c07d97d46b59ea682500887d2a100cda7963f94cc793407ba" exitCode=0 Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.467035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" event={"ID":"3f65e198-57fe-468d-8779-4c25bad24c54","Type":"ContainerDied","Data":"0e8c013c0cd2969c07d97d46b59ea682500887d2a100cda7963f94cc793407ba"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.467417 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" event={"ID":"3f65e198-57fe-468d-8779-4c25bad24c54","Type":"ContainerStarted","Data":"d2e7c6a97477b91b5d6bcd21c1d34e95469a3f2e1441f2ffcde9c6c808678ce4"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.470564 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" event={"ID":"a3871247-7c7c-4ccc-8c58-673cb82b7e97","Type":"ContainerStarted","Data":"aecf76c067a43b4589db15a0213fbe179f4791175d98b5ce1142f1027f021528"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.470613 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" event={"ID":"a3871247-7c7c-4ccc-8c58-673cb82b7e97","Type":"ContainerStarted","Data":"95189c47c5c884c39c5bef3987e180f7a758c7e6dcc648b0e447572e70daeea9"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.470629 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" event={"ID":"a3871247-7c7c-4ccc-8c58-673cb82b7e97","Type":"ContainerStarted","Data":"5a49c473b36c30d18479791095743acf23e6df4d396b6b4d7f9a6a86a869a91f"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.478114 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd5f5578-jr2bf" event={"ID":"ae040e65-3459-4cdc-a35b-e2a47f6bcce3","Type":"ContainerStarted","Data":"c2dacd5b04a38aca6758d40159ba8ca09bb87835aa1f27c6548af5a16e7736f8"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.478167 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd5f5578-jr2bf" event={"ID":"ae040e65-3459-4cdc-a35b-e2a47f6bcce3","Type":"ContainerStarted","Data":"27a314f63b8aae79596004f37cffb22b25863daf1d3c996b6b8b670c3d28c339"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.478181 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd5f5578-jr2bf" event={"ID":"ae040e65-3459-4cdc-a35b-e2a47f6bcce3","Type":"ContainerStarted","Data":"8978a93f90c26e7234ee50d7f5083e72da5098bcc5f0e304f8254939201693b5"} Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.478684 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.478782 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.518404 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-55cd5f5578-jr2bf" podStartSLOduration=2.518385791 podStartE2EDuration="2.518385791s" podCreationTimestamp="2025-12-04 19:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:12:38.515608355 +0000 UTC m=+5620.470969421" watchObservedRunningTime="2025-12-04 19:12:38.518385791 +0000 UTC m=+5620.473746837" Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.567378 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-68d4f54797-jj274" podStartSLOduration=2.5673652049999998 podStartE2EDuration="2.567365205s" podCreationTimestamp="2025-12-04 19:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:12:38.492652825 +0000 UTC m=+5620.448013881" watchObservedRunningTime="2025-12-04 19:12:38.567365205 +0000 UTC m=+5620.522726251" Dec 04 19:12:38 crc kubenswrapper[4733]: I1204 19:12:38.589607 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5bc7fbd458-mz9t9" podStartSLOduration=2.589574676 podStartE2EDuration="2.589574676s" podCreationTimestamp="2025-12-04 19:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:12:38.583958744 +0000 UTC m=+5620.539319790" watchObservedRunningTime="2025-12-04 19:12:38.589574676 +0000 UTC m=+5620.544935722" Dec 04 19:12:39 crc kubenswrapper[4733]: I1204 19:12:39.506625 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" event={"ID":"3f65e198-57fe-468d-8779-4c25bad24c54","Type":"ContainerStarted","Data":"5044dc6f3f176ab12bcfd24b038c9b3e30c41bdfcca17ba0d9543b9ad36c61c1"} Dec 04 19:12:40 crc kubenswrapper[4733]: I1204 19:12:40.517203 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:43 crc kubenswrapper[4733]: I1204 19:12:43.736573 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:43 crc kubenswrapper[4733]: I1204 19:12:43.757702 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" podStartSLOduration=7.757684499 podStartE2EDuration="7.757684499s" podCreationTimestamp="2025-12-04 19:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:12:39.53325676 +0000 UTC m=+5621.488617796" watchObservedRunningTime="2025-12-04 19:12:43.757684499 +0000 UTC m=+5625.713045545" Dec 04 19:12:45 crc kubenswrapper[4733]: I1204 19:12:45.203902 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55cd5f5578-jr2bf" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.143035 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.218473 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f6f5d79c7-b7qkb"] Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.218772 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" podUID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" containerName="dnsmasq-dns" containerID="cri-o://b318c7d579d40af5b01c645d49ad63771afb37831a1ddd4763645127cfa6d4a0" gracePeriod=10 Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.592396 4733 generic.go:334] "Generic (PLEG): container finished" podID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" containerID="b318c7d579d40af5b01c645d49ad63771afb37831a1ddd4763645127cfa6d4a0" exitCode=0 Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.592701 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" event={"ID":"b8cd68b2-cfcb-418d-94c1-588ee8bf6404","Type":"ContainerDied","Data":"b318c7d579d40af5b01c645d49ad63771afb37831a1ddd4763645127cfa6d4a0"} Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.724716 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.833410 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-config\") pod \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.833471 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-dns-svc\") pod \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.833599 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-sb\") pod \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.833681 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-nb\") pod \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.833766 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7zth\" (UniqueName: \"kubernetes.io/projected/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-kube-api-access-k7zth\") pod \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\" (UID: \"b8cd68b2-cfcb-418d-94c1-588ee8bf6404\") " Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.854281 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-kube-api-access-k7zth" (OuterVolumeSpecName: "kube-api-access-k7zth") pod "b8cd68b2-cfcb-418d-94c1-588ee8bf6404" (UID: "b8cd68b2-cfcb-418d-94c1-588ee8bf6404"). InnerVolumeSpecName "kube-api-access-k7zth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.885810 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-config" (OuterVolumeSpecName: "config") pod "b8cd68b2-cfcb-418d-94c1-588ee8bf6404" (UID: "b8cd68b2-cfcb-418d-94c1-588ee8bf6404"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.887270 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b8cd68b2-cfcb-418d-94c1-588ee8bf6404" (UID: "b8cd68b2-cfcb-418d-94c1-588ee8bf6404"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.889320 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b8cd68b2-cfcb-418d-94c1-588ee8bf6404" (UID: "b8cd68b2-cfcb-418d-94c1-588ee8bf6404"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.893705 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b8cd68b2-cfcb-418d-94c1-588ee8bf6404" (UID: "b8cd68b2-cfcb-418d-94c1-588ee8bf6404"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.935837 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.935863 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.935873 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7zth\" (UniqueName: \"kubernetes.io/projected/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-kube-api-access-k7zth\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.935882 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:47 crc kubenswrapper[4733]: I1204 19:12:47.935890 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8cd68b2-cfcb-418d-94c1-588ee8bf6404-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:12:48 crc kubenswrapper[4733]: I1204 19:12:48.602668 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" event={"ID":"b8cd68b2-cfcb-418d-94c1-588ee8bf6404","Type":"ContainerDied","Data":"2dbfc74dd414d066dc900ac5c06a86b57db54df0cb73c1f7f4aa38e3734c8bad"} Dec 04 19:12:48 crc kubenswrapper[4733]: I1204 19:12:48.602727 4733 scope.go:117] "RemoveContainer" containerID="b318c7d579d40af5b01c645d49ad63771afb37831a1ddd4763645127cfa6d4a0" Dec 04 19:12:48 crc kubenswrapper[4733]: I1204 19:12:48.603754 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6f5d79c7-b7qkb" Dec 04 19:12:48 crc kubenswrapper[4733]: I1204 19:12:48.631405 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f6f5d79c7-b7qkb"] Dec 04 19:12:48 crc kubenswrapper[4733]: I1204 19:12:48.632918 4733 scope.go:117] "RemoveContainer" containerID="18d01ddc86637621e9b7db8de514c04f7bac1a444ab855a2498cb7bb5d6ec169" Dec 04 19:12:48 crc kubenswrapper[4733]: I1204 19:12:48.640992 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f6f5d79c7-b7qkb"] Dec 04 19:12:50 crc kubenswrapper[4733]: I1204 19:12:50.344858 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" path="/var/lib/kubelet/pods/b8cd68b2-cfcb-418d-94c1-588ee8bf6404/volumes" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.341495 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-8t6f2"] Dec 04 19:12:57 crc kubenswrapper[4733]: E1204 19:12:57.342438 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" containerName="init" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.342458 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" containerName="init" Dec 04 19:12:57 crc kubenswrapper[4733]: E1204 19:12:57.342477 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" containerName="dnsmasq-dns" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.342484 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" containerName="dnsmasq-dns" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.342689 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8cd68b2-cfcb-418d-94c1-588ee8bf6404" containerName="dnsmasq-dns" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.343455 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8t6f2" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.351300 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8t6f2"] Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.435241 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-53c8-account-create-update-hgqsk"] Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.436304 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.444533 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.448049 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-53c8-account-create-update-hgqsk"] Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.513771 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb8j6\" (UniqueName: \"kubernetes.io/projected/15fb0119-f4e2-4e49-a652-5306d9b0a622-kube-api-access-rb8j6\") pod \"neutron-db-create-8t6f2\" (UID: \"15fb0119-f4e2-4e49-a652-5306d9b0a622\") " pod="openstack/neutron-db-create-8t6f2" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.513913 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15fb0119-f4e2-4e49-a652-5306d9b0a622-operator-scripts\") pod \"neutron-db-create-8t6f2\" (UID: \"15fb0119-f4e2-4e49-a652-5306d9b0a622\") " pod="openstack/neutron-db-create-8t6f2" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.616068 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15fb0119-f4e2-4e49-a652-5306d9b0a622-operator-scripts\") pod \"neutron-db-create-8t6f2\" (UID: \"15fb0119-f4e2-4e49-a652-5306d9b0a622\") " pod="openstack/neutron-db-create-8t6f2" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.616156 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82cpr\" (UniqueName: \"kubernetes.io/projected/31826b3c-ded5-4174-9796-3b8d4aa4abb4-kube-api-access-82cpr\") pod \"neutron-53c8-account-create-update-hgqsk\" (UID: \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\") " pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.616228 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb8j6\" (UniqueName: \"kubernetes.io/projected/15fb0119-f4e2-4e49-a652-5306d9b0a622-kube-api-access-rb8j6\") pod \"neutron-db-create-8t6f2\" (UID: \"15fb0119-f4e2-4e49-a652-5306d9b0a622\") " pod="openstack/neutron-db-create-8t6f2" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.616349 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31826b3c-ded5-4174-9796-3b8d4aa4abb4-operator-scripts\") pod \"neutron-53c8-account-create-update-hgqsk\" (UID: \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\") " pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.617706 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15fb0119-f4e2-4e49-a652-5306d9b0a622-operator-scripts\") pod \"neutron-db-create-8t6f2\" (UID: \"15fb0119-f4e2-4e49-a652-5306d9b0a622\") " pod="openstack/neutron-db-create-8t6f2" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.635396 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb8j6\" (UniqueName: \"kubernetes.io/projected/15fb0119-f4e2-4e49-a652-5306d9b0a622-kube-api-access-rb8j6\") pod \"neutron-db-create-8t6f2\" (UID: \"15fb0119-f4e2-4e49-a652-5306d9b0a622\") " pod="openstack/neutron-db-create-8t6f2" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.663254 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8t6f2" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.717870 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82cpr\" (UniqueName: \"kubernetes.io/projected/31826b3c-ded5-4174-9796-3b8d4aa4abb4-kube-api-access-82cpr\") pod \"neutron-53c8-account-create-update-hgqsk\" (UID: \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\") " pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.718203 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31826b3c-ded5-4174-9796-3b8d4aa4abb4-operator-scripts\") pod \"neutron-53c8-account-create-update-hgqsk\" (UID: \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\") " pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.719510 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31826b3c-ded5-4174-9796-3b8d4aa4abb4-operator-scripts\") pod \"neutron-53c8-account-create-update-hgqsk\" (UID: \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\") " pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.744602 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82cpr\" (UniqueName: \"kubernetes.io/projected/31826b3c-ded5-4174-9796-3b8d4aa4abb4-kube-api-access-82cpr\") pod \"neutron-53c8-account-create-update-hgqsk\" (UID: \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\") " pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:12:57 crc kubenswrapper[4733]: I1204 19:12:57.751878 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:12:58 crc kubenswrapper[4733]: I1204 19:12:58.134090 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8t6f2"] Dec 04 19:12:58 crc kubenswrapper[4733]: W1204 19:12:58.140424 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15fb0119_f4e2_4e49_a652_5306d9b0a622.slice/crio-1ce6cf95e3b793d0d69df62630a60e6c6962e9e18bd1aea7e9436214afad6afa WatchSource:0}: Error finding container 1ce6cf95e3b793d0d69df62630a60e6c6962e9e18bd1aea7e9436214afad6afa: Status 404 returned error can't find the container with id 1ce6cf95e3b793d0d69df62630a60e6c6962e9e18bd1aea7e9436214afad6afa Dec 04 19:12:58 crc kubenswrapper[4733]: I1204 19:12:58.218196 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-53c8-account-create-update-hgqsk"] Dec 04 19:12:58 crc kubenswrapper[4733]: W1204 19:12:58.219896 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31826b3c_ded5_4174_9796_3b8d4aa4abb4.slice/crio-983a56ebb9110f013f52d0b80557635ceab68b93035f5989f6a1f52fadd254b9 WatchSource:0}: Error finding container 983a56ebb9110f013f52d0b80557635ceab68b93035f5989f6a1f52fadd254b9: Status 404 returned error can't find the container with id 983a56ebb9110f013f52d0b80557635ceab68b93035f5989f6a1f52fadd254b9 Dec 04 19:12:58 crc kubenswrapper[4733]: I1204 19:12:58.738654 4733 generic.go:334] "Generic (PLEG): container finished" podID="31826b3c-ded5-4174-9796-3b8d4aa4abb4" containerID="3669ce59afc64860ea6608339208ab3b75b9c7fcb25f0e1d5c3c83f760b4eea9" exitCode=0 Dec 04 19:12:58 crc kubenswrapper[4733]: I1204 19:12:58.738856 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-53c8-account-create-update-hgqsk" event={"ID":"31826b3c-ded5-4174-9796-3b8d4aa4abb4","Type":"ContainerDied","Data":"3669ce59afc64860ea6608339208ab3b75b9c7fcb25f0e1d5c3c83f760b4eea9"} Dec 04 19:12:58 crc kubenswrapper[4733]: I1204 19:12:58.738926 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-53c8-account-create-update-hgqsk" event={"ID":"31826b3c-ded5-4174-9796-3b8d4aa4abb4","Type":"ContainerStarted","Data":"983a56ebb9110f013f52d0b80557635ceab68b93035f5989f6a1f52fadd254b9"} Dec 04 19:12:58 crc kubenswrapper[4733]: I1204 19:12:58.741507 4733 generic.go:334] "Generic (PLEG): container finished" podID="15fb0119-f4e2-4e49-a652-5306d9b0a622" containerID="9b1ada6de779a7c2d39facc56c0642fe857486bc6e61159cd487d610f5ed8ed3" exitCode=0 Dec 04 19:12:58 crc kubenswrapper[4733]: I1204 19:12:58.741553 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8t6f2" event={"ID":"15fb0119-f4e2-4e49-a652-5306d9b0a622","Type":"ContainerDied","Data":"9b1ada6de779a7c2d39facc56c0642fe857486bc6e61159cd487d610f5ed8ed3"} Dec 04 19:12:58 crc kubenswrapper[4733]: I1204 19:12:58.741594 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8t6f2" event={"ID":"15fb0119-f4e2-4e49-a652-5306d9b0a622","Type":"ContainerStarted","Data":"1ce6cf95e3b793d0d69df62630a60e6c6962e9e18bd1aea7e9436214afad6afa"} Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.225135 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.231727 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8t6f2" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.368426 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb8j6\" (UniqueName: \"kubernetes.io/projected/15fb0119-f4e2-4e49-a652-5306d9b0a622-kube-api-access-rb8j6\") pod \"15fb0119-f4e2-4e49-a652-5306d9b0a622\" (UID: \"15fb0119-f4e2-4e49-a652-5306d9b0a622\") " Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.368765 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82cpr\" (UniqueName: \"kubernetes.io/projected/31826b3c-ded5-4174-9796-3b8d4aa4abb4-kube-api-access-82cpr\") pod \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\" (UID: \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\") " Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.368971 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31826b3c-ded5-4174-9796-3b8d4aa4abb4-operator-scripts\") pod \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\" (UID: \"31826b3c-ded5-4174-9796-3b8d4aa4abb4\") " Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.369047 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15fb0119-f4e2-4e49-a652-5306d9b0a622-operator-scripts\") pod \"15fb0119-f4e2-4e49-a652-5306d9b0a622\" (UID: \"15fb0119-f4e2-4e49-a652-5306d9b0a622\") " Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.370656 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31826b3c-ded5-4174-9796-3b8d4aa4abb4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31826b3c-ded5-4174-9796-3b8d4aa4abb4" (UID: "31826b3c-ded5-4174-9796-3b8d4aa4abb4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.370872 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15fb0119-f4e2-4e49-a652-5306d9b0a622-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15fb0119-f4e2-4e49-a652-5306d9b0a622" (UID: "15fb0119-f4e2-4e49-a652-5306d9b0a622"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.377922 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15fb0119-f4e2-4e49-a652-5306d9b0a622-kube-api-access-rb8j6" (OuterVolumeSpecName: "kube-api-access-rb8j6") pod "15fb0119-f4e2-4e49-a652-5306d9b0a622" (UID: "15fb0119-f4e2-4e49-a652-5306d9b0a622"). InnerVolumeSpecName "kube-api-access-rb8j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.378092 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31826b3c-ded5-4174-9796-3b8d4aa4abb4-kube-api-access-82cpr" (OuterVolumeSpecName: "kube-api-access-82cpr") pod "31826b3c-ded5-4174-9796-3b8d4aa4abb4" (UID: "31826b3c-ded5-4174-9796-3b8d4aa4abb4"). InnerVolumeSpecName "kube-api-access-82cpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.472847 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb8j6\" (UniqueName: \"kubernetes.io/projected/15fb0119-f4e2-4e49-a652-5306d9b0a622-kube-api-access-rb8j6\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.472901 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82cpr\" (UniqueName: \"kubernetes.io/projected/31826b3c-ded5-4174-9796-3b8d4aa4abb4-kube-api-access-82cpr\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.472922 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31826b3c-ded5-4174-9796-3b8d4aa4abb4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.472939 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15fb0119-f4e2-4e49-a652-5306d9b0a622-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.764242 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8t6f2" event={"ID":"15fb0119-f4e2-4e49-a652-5306d9b0a622","Type":"ContainerDied","Data":"1ce6cf95e3b793d0d69df62630a60e6c6962e9e18bd1aea7e9436214afad6afa"} Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.765097 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ce6cf95e3b793d0d69df62630a60e6c6962e9e18bd1aea7e9436214afad6afa" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.764300 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8t6f2" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.767299 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-53c8-account-create-update-hgqsk" event={"ID":"31826b3c-ded5-4174-9796-3b8d4aa4abb4","Type":"ContainerDied","Data":"983a56ebb9110f013f52d0b80557635ceab68b93035f5989f6a1f52fadd254b9"} Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.767365 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="983a56ebb9110f013f52d0b80557635ceab68b93035f5989f6a1f52fadd254b9" Dec 04 19:13:00 crc kubenswrapper[4733]: I1204 19:13:00.767512 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-53c8-account-create-update-hgqsk" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.744137 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2vgfw"] Dec 04 19:13:02 crc kubenswrapper[4733]: E1204 19:13:02.744963 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31826b3c-ded5-4174-9796-3b8d4aa4abb4" containerName="mariadb-account-create-update" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.744984 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="31826b3c-ded5-4174-9796-3b8d4aa4abb4" containerName="mariadb-account-create-update" Dec 04 19:13:02 crc kubenswrapper[4733]: E1204 19:13:02.745006 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15fb0119-f4e2-4e49-a652-5306d9b0a622" containerName="mariadb-database-create" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.745017 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="15fb0119-f4e2-4e49-a652-5306d9b0a622" containerName="mariadb-database-create" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.745317 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="31826b3c-ded5-4174-9796-3b8d4aa4abb4" containerName="mariadb-account-create-update" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.745357 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="15fb0119-f4e2-4e49-a652-5306d9b0a622" containerName="mariadb-database-create" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.746194 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.748207 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-7m9lm" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.748210 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.749556 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.762879 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2vgfw"] Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.917599 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96cvm\" (UniqueName: \"kubernetes.io/projected/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-kube-api-access-96cvm\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.917703 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-config\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:02 crc kubenswrapper[4733]: I1204 19:13:02.917882 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-combined-ca-bundle\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.019123 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96cvm\" (UniqueName: \"kubernetes.io/projected/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-kube-api-access-96cvm\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.019176 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-config\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.019204 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-combined-ca-bundle\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.025755 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-combined-ca-bundle\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.031629 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-config\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.036635 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96cvm\" (UniqueName: \"kubernetes.io/projected/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-kube-api-access-96cvm\") pod \"neutron-db-sync-2vgfw\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.066474 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.559623 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2vgfw"] Dec 04 19:13:03 crc kubenswrapper[4733]: I1204 19:13:03.799282 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2vgfw" event={"ID":"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77","Type":"ContainerStarted","Data":"bdea2b4c2214933b0497ec5d1704de696817f9541fffee1e3527fd52983caec8"} Dec 04 19:13:04 crc kubenswrapper[4733]: I1204 19:13:04.818451 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2vgfw" event={"ID":"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77","Type":"ContainerStarted","Data":"98a4162d8e1bd916f1f76ec054ae8304f012c17dbc2741a3a956ef56f37882a8"} Dec 04 19:13:04 crc kubenswrapper[4733]: I1204 19:13:04.836590 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2vgfw" podStartSLOduration=2.836570519 podStartE2EDuration="2.836570519s" podCreationTimestamp="2025-12-04 19:13:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:13:04.834450731 +0000 UTC m=+5646.789811817" watchObservedRunningTime="2025-12-04 19:13:04.836570519 +0000 UTC m=+5646.791931565" Dec 04 19:13:07 crc kubenswrapper[4733]: I1204 19:13:07.864826 4733 generic.go:334] "Generic (PLEG): container finished" podID="eaaf0141-6b6c-4e56-8fc1-c7a875c3be77" containerID="98a4162d8e1bd916f1f76ec054ae8304f012c17dbc2741a3a956ef56f37882a8" exitCode=0 Dec 04 19:13:07 crc kubenswrapper[4733]: I1204 19:13:07.864875 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2vgfw" event={"ID":"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77","Type":"ContainerDied","Data":"98a4162d8e1bd916f1f76ec054ae8304f012c17dbc2741a3a956ef56f37882a8"} Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.286000 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.341068 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96cvm\" (UniqueName: \"kubernetes.io/projected/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-kube-api-access-96cvm\") pod \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.341125 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-combined-ca-bundle\") pod \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.341245 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-config\") pod \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\" (UID: \"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77\") " Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.349373 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-kube-api-access-96cvm" (OuterVolumeSpecName: "kube-api-access-96cvm") pod "eaaf0141-6b6c-4e56-8fc1-c7a875c3be77" (UID: "eaaf0141-6b6c-4e56-8fc1-c7a875c3be77"). InnerVolumeSpecName "kube-api-access-96cvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.365124 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-config" (OuterVolumeSpecName: "config") pod "eaaf0141-6b6c-4e56-8fc1-c7a875c3be77" (UID: "eaaf0141-6b6c-4e56-8fc1-c7a875c3be77"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.382370 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaaf0141-6b6c-4e56-8fc1-c7a875c3be77" (UID: "eaaf0141-6b6c-4e56-8fc1-c7a875c3be77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.443896 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.443956 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96cvm\" (UniqueName: \"kubernetes.io/projected/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-kube-api-access-96cvm\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.443994 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.887869 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2vgfw" event={"ID":"eaaf0141-6b6c-4e56-8fc1-c7a875c3be77","Type":"ContainerDied","Data":"bdea2b4c2214933b0497ec5d1704de696817f9541fffee1e3527fd52983caec8"} Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.887927 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdea2b4c2214933b0497ec5d1704de696817f9541fffee1e3527fd52983caec8" Dec 04 19:13:09 crc kubenswrapper[4733]: I1204 19:13:09.888006 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2vgfw" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.159284 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d466c8f4c-vn5n2"] Dec 04 19:13:10 crc kubenswrapper[4733]: E1204 19:13:10.159984 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaaf0141-6b6c-4e56-8fc1-c7a875c3be77" containerName="neutron-db-sync" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.160005 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaaf0141-6b6c-4e56-8fc1-c7a875c3be77" containerName="neutron-db-sync" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.160264 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaaf0141-6b6c-4e56-8fc1-c7a875c3be77" containerName="neutron-db-sync" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.161385 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.175291 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d466c8f4c-vn5n2"] Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.257763 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-nb\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.258034 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-sb\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.258125 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx49v\" (UniqueName: \"kubernetes.io/projected/566dedb2-20cf-44a9-97f0-47f19706d36a-kube-api-access-fx49v\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.258238 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-config\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.258327 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-dns-svc\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.360008 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-nb\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.360086 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-sb\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.360116 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx49v\" (UniqueName: \"kubernetes.io/projected/566dedb2-20cf-44a9-97f0-47f19706d36a-kube-api-access-fx49v\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.360185 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-config\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.360219 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-dns-svc\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.361330 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-dns-svc\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.361395 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-nb\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.361464 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-sb\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.361712 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-config\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.376078 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-67fb4dd6bf-g6962"] Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.377463 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.380131 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.380315 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-7m9lm" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.380482 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.385972 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx49v\" (UniqueName: \"kubernetes.io/projected/566dedb2-20cf-44a9-97f0-47f19706d36a-kube-api-access-fx49v\") pod \"dnsmasq-dns-6d466c8f4c-vn5n2\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.393250 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67fb4dd6bf-g6962"] Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.461936 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-combined-ca-bundle\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.462023 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-config\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.462096 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-httpd-config\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.462169 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx74t\" (UniqueName: \"kubernetes.io/projected/e23aec62-3331-416d-868e-a4ed6b471984-kube-api-access-dx74t\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.481650 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.563676 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-config\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.564369 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-httpd-config\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.564540 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx74t\" (UniqueName: \"kubernetes.io/projected/e23aec62-3331-416d-868e-a4ed6b471984-kube-api-access-dx74t\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.564699 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-combined-ca-bundle\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.575274 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-combined-ca-bundle\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.576053 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-config\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.576328 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e23aec62-3331-416d-868e-a4ed6b471984-httpd-config\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.583062 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx74t\" (UniqueName: \"kubernetes.io/projected/e23aec62-3331-416d-868e-a4ed6b471984-kube-api-access-dx74t\") pod \"neutron-67fb4dd6bf-g6962\" (UID: \"e23aec62-3331-416d-868e-a4ed6b471984\") " pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.742821 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:10 crc kubenswrapper[4733]: I1204 19:13:10.984004 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d466c8f4c-vn5n2"] Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.339866 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67fb4dd6bf-g6962"] Dec 04 19:13:11 crc kubenswrapper[4733]: W1204 19:13:11.344271 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode23aec62_3331_416d_868e_a4ed6b471984.slice/crio-f564d8c78dbdeb2903a8bd1dafd4b7a12a6c01dd3d0d1611e13bedfcebcfaa77 WatchSource:0}: Error finding container f564d8c78dbdeb2903a8bd1dafd4b7a12a6c01dd3d0d1611e13bedfcebcfaa77: Status 404 returned error can't find the container with id f564d8c78dbdeb2903a8bd1dafd4b7a12a6c01dd3d0d1611e13bedfcebcfaa77 Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.905219 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67fb4dd6bf-g6962" event={"ID":"e23aec62-3331-416d-868e-a4ed6b471984","Type":"ContainerStarted","Data":"c05200ca2bdb4fe93cd9deff4f5059a302cbc1a05b9e2fb45d7d799a50295a13"} Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.905577 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67fb4dd6bf-g6962" event={"ID":"e23aec62-3331-416d-868e-a4ed6b471984","Type":"ContainerStarted","Data":"d138d3b485a559cd52dd87dd249c7cdd2934156a9137363b467d565bd1867703"} Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.905594 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67fb4dd6bf-g6962" event={"ID":"e23aec62-3331-416d-868e-a4ed6b471984","Type":"ContainerStarted","Data":"f564d8c78dbdeb2903a8bd1dafd4b7a12a6c01dd3d0d1611e13bedfcebcfaa77"} Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.905614 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.908019 4733 generic.go:334] "Generic (PLEG): container finished" podID="566dedb2-20cf-44a9-97f0-47f19706d36a" containerID="851d082c84fcb82e2e712fc926f773f7c91249c843d15956f59f583ff1f4d392" exitCode=0 Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.908066 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" event={"ID":"566dedb2-20cf-44a9-97f0-47f19706d36a","Type":"ContainerDied","Data":"851d082c84fcb82e2e712fc926f773f7c91249c843d15956f59f583ff1f4d392"} Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.908098 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" event={"ID":"566dedb2-20cf-44a9-97f0-47f19706d36a","Type":"ContainerStarted","Data":"eca9d77885a1d27190003ac1b9c4c02fbc0c1429927487326b550f7abc5b08c2"} Dec 04 19:13:11 crc kubenswrapper[4733]: I1204 19:13:11.927861 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-67fb4dd6bf-g6962" podStartSLOduration=1.927843725 podStartE2EDuration="1.927843725s" podCreationTimestamp="2025-12-04 19:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:13:11.920307541 +0000 UTC m=+5653.875668587" watchObservedRunningTime="2025-12-04 19:13:11.927843725 +0000 UTC m=+5653.883204761" Dec 04 19:13:12 crc kubenswrapper[4733]: I1204 19:13:12.919963 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" event={"ID":"566dedb2-20cf-44a9-97f0-47f19706d36a","Type":"ContainerStarted","Data":"25f1247a2aecf9686eff30239023b84a3152d22e84c36bafed5de8505c33ce34"} Dec 04 19:13:12 crc kubenswrapper[4733]: I1204 19:13:12.920348 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:12 crc kubenswrapper[4733]: I1204 19:13:12.959641 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" podStartSLOduration=2.959606771 podStartE2EDuration="2.959606771s" podCreationTimestamp="2025-12-04 19:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:13:12.946451846 +0000 UTC m=+5654.901812902" watchObservedRunningTime="2025-12-04 19:13:12.959606771 +0000 UTC m=+5654.914967857" Dec 04 19:13:20 crc kubenswrapper[4733]: I1204 19:13:20.484007 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:13:20 crc kubenswrapper[4733]: I1204 19:13:20.549469 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dfdb7c58c-vpkhw"] Dec 04 19:13:20 crc kubenswrapper[4733]: I1204 19:13:20.549692 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" podUID="3f65e198-57fe-468d-8779-4c25bad24c54" containerName="dnsmasq-dns" containerID="cri-o://5044dc6f3f176ab12bcfd24b038c9b3e30c41bdfcca17ba0d9543b9ad36c61c1" gracePeriod=10 Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.006591 4733 generic.go:334] "Generic (PLEG): container finished" podID="3f65e198-57fe-468d-8779-4c25bad24c54" containerID="5044dc6f3f176ab12bcfd24b038c9b3e30c41bdfcca17ba0d9543b9ad36c61c1" exitCode=0 Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.006928 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" event={"ID":"3f65e198-57fe-468d-8779-4c25bad24c54","Type":"ContainerDied","Data":"5044dc6f3f176ab12bcfd24b038c9b3e30c41bdfcca17ba0d9543b9ad36c61c1"} Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.007029 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" event={"ID":"3f65e198-57fe-468d-8779-4c25bad24c54","Type":"ContainerDied","Data":"d2e7c6a97477b91b5d6bcd21c1d34e95469a3f2e1441f2ffcde9c6c808678ce4"} Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.007041 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2e7c6a97477b91b5d6bcd21c1d34e95469a3f2e1441f2ffcde9c6c808678ce4" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.085828 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.150486 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-nb\") pod \"3f65e198-57fe-468d-8779-4c25bad24c54\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.150584 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-sb\") pod \"3f65e198-57fe-468d-8779-4c25bad24c54\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.150636 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-dns-svc\") pod \"3f65e198-57fe-468d-8779-4c25bad24c54\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.150680 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-config\") pod \"3f65e198-57fe-468d-8779-4c25bad24c54\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.151346 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms8cq\" (UniqueName: \"kubernetes.io/projected/3f65e198-57fe-468d-8779-4c25bad24c54-kube-api-access-ms8cq\") pod \"3f65e198-57fe-468d-8779-4c25bad24c54\" (UID: \"3f65e198-57fe-468d-8779-4c25bad24c54\") " Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.251748 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f65e198-57fe-468d-8779-4c25bad24c54-kube-api-access-ms8cq" (OuterVolumeSpecName: "kube-api-access-ms8cq") pod "3f65e198-57fe-468d-8779-4c25bad24c54" (UID: "3f65e198-57fe-468d-8779-4c25bad24c54"). InnerVolumeSpecName "kube-api-access-ms8cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.253461 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms8cq\" (UniqueName: \"kubernetes.io/projected/3f65e198-57fe-468d-8779-4c25bad24c54-kube-api-access-ms8cq\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.273277 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-config" (OuterVolumeSpecName: "config") pod "3f65e198-57fe-468d-8779-4c25bad24c54" (UID: "3f65e198-57fe-468d-8779-4c25bad24c54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.276194 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3f65e198-57fe-468d-8779-4c25bad24c54" (UID: "3f65e198-57fe-468d-8779-4c25bad24c54"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.276487 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3f65e198-57fe-468d-8779-4c25bad24c54" (UID: "3f65e198-57fe-468d-8779-4c25bad24c54"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.279331 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3f65e198-57fe-468d-8779-4c25bad24c54" (UID: "3f65e198-57fe-468d-8779-4c25bad24c54"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.355026 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.355089 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.355103 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:21 crc kubenswrapper[4733]: I1204 19:13:21.355115 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f65e198-57fe-468d-8779-4c25bad24c54-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:22 crc kubenswrapper[4733]: I1204 19:13:22.014394 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfdb7c58c-vpkhw" Dec 04 19:13:22 crc kubenswrapper[4733]: I1204 19:13:22.053169 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dfdb7c58c-vpkhw"] Dec 04 19:13:22 crc kubenswrapper[4733]: I1204 19:13:22.061465 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dfdb7c58c-vpkhw"] Dec 04 19:13:22 crc kubenswrapper[4733]: I1204 19:13:22.356097 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f65e198-57fe-468d-8779-4c25bad24c54" path="/var/lib/kubelet/pods/3f65e198-57fe-468d-8779-4c25bad24c54/volumes" Dec 04 19:13:40 crc kubenswrapper[4733]: I1204 19:13:40.761598 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-67fb4dd6bf-g6962" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.614712 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-zf7qx"] Dec 04 19:13:47 crc kubenswrapper[4733]: E1204 19:13:47.615969 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f65e198-57fe-468d-8779-4c25bad24c54" containerName="dnsmasq-dns" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.615994 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f65e198-57fe-468d-8779-4c25bad24c54" containerName="dnsmasq-dns" Dec 04 19:13:47 crc kubenswrapper[4733]: E1204 19:13:47.616020 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f65e198-57fe-468d-8779-4c25bad24c54" containerName="init" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.616034 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f65e198-57fe-468d-8779-4c25bad24c54" containerName="init" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.616380 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f65e198-57fe-468d-8779-4c25bad24c54" containerName="dnsmasq-dns" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.619177 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.628003 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zf7qx"] Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.711753 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b55f-account-create-update-5mslq"] Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.712926 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd069278-c6ab-4341-a9ed-1b93d4aa6280-operator-scripts\") pod \"glance-db-create-zf7qx\" (UID: \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\") " pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.720135 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf6x4\" (UniqueName: \"kubernetes.io/projected/dd069278-c6ab-4341-a9ed-1b93d4aa6280-kube-api-access-gf6x4\") pod \"glance-db-create-zf7qx\" (UID: \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\") " pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.715523 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.720843 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b55f-account-create-update-5mslq"] Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.722158 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.821935 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd069278-c6ab-4341-a9ed-1b93d4aa6280-operator-scripts\") pod \"glance-db-create-zf7qx\" (UID: \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\") " pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.822012 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e986a553-d439-4add-90ea-a3d15d378c29-operator-scripts\") pod \"glance-b55f-account-create-update-5mslq\" (UID: \"e986a553-d439-4add-90ea-a3d15d378c29\") " pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.822057 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf6x4\" (UniqueName: \"kubernetes.io/projected/dd069278-c6ab-4341-a9ed-1b93d4aa6280-kube-api-access-gf6x4\") pod \"glance-db-create-zf7qx\" (UID: \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\") " pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.822089 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqlqb\" (UniqueName: \"kubernetes.io/projected/e986a553-d439-4add-90ea-a3d15d378c29-kube-api-access-gqlqb\") pod \"glance-b55f-account-create-update-5mslq\" (UID: \"e986a553-d439-4add-90ea-a3d15d378c29\") " pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.822946 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd069278-c6ab-4341-a9ed-1b93d4aa6280-operator-scripts\") pod \"glance-db-create-zf7qx\" (UID: \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\") " pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.838967 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf6x4\" (UniqueName: \"kubernetes.io/projected/dd069278-c6ab-4341-a9ed-1b93d4aa6280-kube-api-access-gf6x4\") pod \"glance-db-create-zf7qx\" (UID: \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\") " pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.923803 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e986a553-d439-4add-90ea-a3d15d378c29-operator-scripts\") pod \"glance-b55f-account-create-update-5mslq\" (UID: \"e986a553-d439-4add-90ea-a3d15d378c29\") " pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.923875 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqlqb\" (UniqueName: \"kubernetes.io/projected/e986a553-d439-4add-90ea-a3d15d378c29-kube-api-access-gqlqb\") pod \"glance-b55f-account-create-update-5mslq\" (UID: \"e986a553-d439-4add-90ea-a3d15d378c29\") " pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.924722 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e986a553-d439-4add-90ea-a3d15d378c29-operator-scripts\") pod \"glance-b55f-account-create-update-5mslq\" (UID: \"e986a553-d439-4add-90ea-a3d15d378c29\") " pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.952390 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqlqb\" (UniqueName: \"kubernetes.io/projected/e986a553-d439-4add-90ea-a3d15d378c29-kube-api-access-gqlqb\") pod \"glance-b55f-account-create-update-5mslq\" (UID: \"e986a553-d439-4add-90ea-a3d15d378c29\") " pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:47 crc kubenswrapper[4733]: I1204 19:13:47.957528 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:48 crc kubenswrapper[4733]: I1204 19:13:48.038213 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:48 crc kubenswrapper[4733]: I1204 19:13:48.473183 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zf7qx"] Dec 04 19:13:48 crc kubenswrapper[4733]: W1204 19:13:48.481505 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd069278_c6ab_4341_a9ed_1b93d4aa6280.slice/crio-78a06172e391bce86866bb1e85e0cf08a8e05e278c8450e9d0d3a1b02ec82008 WatchSource:0}: Error finding container 78a06172e391bce86866bb1e85e0cf08a8e05e278c8450e9d0d3a1b02ec82008: Status 404 returned error can't find the container with id 78a06172e391bce86866bb1e85e0cf08a8e05e278c8450e9d0d3a1b02ec82008 Dec 04 19:13:48 crc kubenswrapper[4733]: I1204 19:13:48.552626 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b55f-account-create-update-5mslq"] Dec 04 19:13:49 crc kubenswrapper[4733]: I1204 19:13:49.260068 4733 generic.go:334] "Generic (PLEG): container finished" podID="dd069278-c6ab-4341-a9ed-1b93d4aa6280" containerID="30be18b9098e9150bac2d3e92ff313a66b5f42c5c7df82fcd9b06eac20686d81" exitCode=0 Dec 04 19:13:49 crc kubenswrapper[4733]: I1204 19:13:49.260184 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zf7qx" event={"ID":"dd069278-c6ab-4341-a9ed-1b93d4aa6280","Type":"ContainerDied","Data":"30be18b9098e9150bac2d3e92ff313a66b5f42c5c7df82fcd9b06eac20686d81"} Dec 04 19:13:49 crc kubenswrapper[4733]: I1204 19:13:49.260500 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zf7qx" event={"ID":"dd069278-c6ab-4341-a9ed-1b93d4aa6280","Type":"ContainerStarted","Data":"78a06172e391bce86866bb1e85e0cf08a8e05e278c8450e9d0d3a1b02ec82008"} Dec 04 19:13:49 crc kubenswrapper[4733]: I1204 19:13:49.262647 4733 generic.go:334] "Generic (PLEG): container finished" podID="e986a553-d439-4add-90ea-a3d15d378c29" containerID="d86ee3847eee51d74f52491ec6b66e0b770d53015dd77d5e3e38986499bafac9" exitCode=0 Dec 04 19:13:49 crc kubenswrapper[4733]: I1204 19:13:49.262702 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b55f-account-create-update-5mslq" event={"ID":"e986a553-d439-4add-90ea-a3d15d378c29","Type":"ContainerDied","Data":"d86ee3847eee51d74f52491ec6b66e0b770d53015dd77d5e3e38986499bafac9"} Dec 04 19:13:49 crc kubenswrapper[4733]: I1204 19:13:49.262725 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b55f-account-create-update-5mslq" event={"ID":"e986a553-d439-4add-90ea-a3d15d378c29","Type":"ContainerStarted","Data":"5d7e7c5482351be3905ac11377be71ba6532595822eb2ae2a2fe7a7c8645d5ad"} Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.681966 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.687776 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.787593 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf6x4\" (UniqueName: \"kubernetes.io/projected/dd069278-c6ab-4341-a9ed-1b93d4aa6280-kube-api-access-gf6x4\") pod \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\" (UID: \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\") " Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.787891 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e986a553-d439-4add-90ea-a3d15d378c29-operator-scripts\") pod \"e986a553-d439-4add-90ea-a3d15d378c29\" (UID: \"e986a553-d439-4add-90ea-a3d15d378c29\") " Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.787933 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd069278-c6ab-4341-a9ed-1b93d4aa6280-operator-scripts\") pod \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\" (UID: \"dd069278-c6ab-4341-a9ed-1b93d4aa6280\") " Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.788904 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e986a553-d439-4add-90ea-a3d15d378c29-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e986a553-d439-4add-90ea-a3d15d378c29" (UID: "e986a553-d439-4add-90ea-a3d15d378c29"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.788904 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd069278-c6ab-4341-a9ed-1b93d4aa6280-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd069278-c6ab-4341-a9ed-1b93d4aa6280" (UID: "dd069278-c6ab-4341-a9ed-1b93d4aa6280"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.789008 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqlqb\" (UniqueName: \"kubernetes.io/projected/e986a553-d439-4add-90ea-a3d15d378c29-kube-api-access-gqlqb\") pod \"e986a553-d439-4add-90ea-a3d15d378c29\" (UID: \"e986a553-d439-4add-90ea-a3d15d378c29\") " Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.789884 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e986a553-d439-4add-90ea-a3d15d378c29-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.789918 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd069278-c6ab-4341-a9ed-1b93d4aa6280-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.794330 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e986a553-d439-4add-90ea-a3d15d378c29-kube-api-access-gqlqb" (OuterVolumeSpecName: "kube-api-access-gqlqb") pod "e986a553-d439-4add-90ea-a3d15d378c29" (UID: "e986a553-d439-4add-90ea-a3d15d378c29"). InnerVolumeSpecName "kube-api-access-gqlqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.796378 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd069278-c6ab-4341-a9ed-1b93d4aa6280-kube-api-access-gf6x4" (OuterVolumeSpecName: "kube-api-access-gf6x4") pod "dd069278-c6ab-4341-a9ed-1b93d4aa6280" (UID: "dd069278-c6ab-4341-a9ed-1b93d4aa6280"). InnerVolumeSpecName "kube-api-access-gf6x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.891605 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqlqb\" (UniqueName: \"kubernetes.io/projected/e986a553-d439-4add-90ea-a3d15d378c29-kube-api-access-gqlqb\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:50 crc kubenswrapper[4733]: I1204 19:13:50.891646 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf6x4\" (UniqueName: \"kubernetes.io/projected/dd069278-c6ab-4341-a9ed-1b93d4aa6280-kube-api-access-gf6x4\") on node \"crc\" DevicePath \"\"" Dec 04 19:13:51 crc kubenswrapper[4733]: I1204 19:13:51.282880 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b55f-account-create-update-5mslq" event={"ID":"e986a553-d439-4add-90ea-a3d15d378c29","Type":"ContainerDied","Data":"5d7e7c5482351be3905ac11377be71ba6532595822eb2ae2a2fe7a7c8645d5ad"} Dec 04 19:13:51 crc kubenswrapper[4733]: I1204 19:13:51.282921 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d7e7c5482351be3905ac11377be71ba6532595822eb2ae2a2fe7a7c8645d5ad" Dec 04 19:13:51 crc kubenswrapper[4733]: I1204 19:13:51.282987 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b55f-account-create-update-5mslq" Dec 04 19:13:51 crc kubenswrapper[4733]: I1204 19:13:51.285385 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zf7qx" event={"ID":"dd069278-c6ab-4341-a9ed-1b93d4aa6280","Type":"ContainerDied","Data":"78a06172e391bce86866bb1e85e0cf08a8e05e278c8450e9d0d3a1b02ec82008"} Dec 04 19:13:51 crc kubenswrapper[4733]: I1204 19:13:51.285423 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78a06172e391bce86866bb1e85e0cf08a8e05e278c8450e9d0d3a1b02ec82008" Dec 04 19:13:51 crc kubenswrapper[4733]: I1204 19:13:51.285476 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zf7qx" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.946490 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-mgml8"] Dec 04 19:13:52 crc kubenswrapper[4733]: E1204 19:13:52.946917 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd069278-c6ab-4341-a9ed-1b93d4aa6280" containerName="mariadb-database-create" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.946933 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd069278-c6ab-4341-a9ed-1b93d4aa6280" containerName="mariadb-database-create" Dec 04 19:13:52 crc kubenswrapper[4733]: E1204 19:13:52.946979 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e986a553-d439-4add-90ea-a3d15d378c29" containerName="mariadb-account-create-update" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.946987 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e986a553-d439-4add-90ea-a3d15d378c29" containerName="mariadb-account-create-update" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.947179 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e986a553-d439-4add-90ea-a3d15d378c29" containerName="mariadb-account-create-update" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.947196 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd069278-c6ab-4341-a9ed-1b93d4aa6280" containerName="mariadb-database-create" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.947989 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.950536 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.950686 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-r76xm" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.966602 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mgml8"] Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.967335 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms4xz\" (UniqueName: \"kubernetes.io/projected/0d88cdb3-e89d-4743-a918-4485083caaf5-kube-api-access-ms4xz\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.967420 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-config-data\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.967555 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-combined-ca-bundle\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:52 crc kubenswrapper[4733]: I1204 19:13:52.967643 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-db-sync-config-data\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.068989 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms4xz\" (UniqueName: \"kubernetes.io/projected/0d88cdb3-e89d-4743-a918-4485083caaf5-kube-api-access-ms4xz\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.069364 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-config-data\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.069407 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-combined-ca-bundle\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.069443 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-db-sync-config-data\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.073823 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-combined-ca-bundle\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.085244 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-db-sync-config-data\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.086063 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms4xz\" (UniqueName: \"kubernetes.io/projected/0d88cdb3-e89d-4743-a918-4485083caaf5-kube-api-access-ms4xz\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.086382 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-config-data\") pod \"glance-db-sync-mgml8\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.265040 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mgml8" Dec 04 19:13:53 crc kubenswrapper[4733]: I1204 19:13:53.921537 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mgml8"] Dec 04 19:13:53 crc kubenswrapper[4733]: W1204 19:13:53.932708 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d88cdb3_e89d_4743_a918_4485083caaf5.slice/crio-928ec252c6e281f72fba35ffb9e591070b6949cb0a4ee4059f76fc4c927ac23e WatchSource:0}: Error finding container 928ec252c6e281f72fba35ffb9e591070b6949cb0a4ee4059f76fc4c927ac23e: Status 404 returned error can't find the container with id 928ec252c6e281f72fba35ffb9e591070b6949cb0a4ee4059f76fc4c927ac23e Dec 04 19:13:54 crc kubenswrapper[4733]: I1204 19:13:54.374956 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mgml8" event={"ID":"0d88cdb3-e89d-4743-a918-4485083caaf5","Type":"ContainerStarted","Data":"928ec252c6e281f72fba35ffb9e591070b6949cb0a4ee4059f76fc4c927ac23e"} Dec 04 19:13:55 crc kubenswrapper[4733]: I1204 19:13:55.385986 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mgml8" event={"ID":"0d88cdb3-e89d-4743-a918-4485083caaf5","Type":"ContainerStarted","Data":"a60c6dc6ddfd19c65af9c8a9b2d5c196b947bfe4a4e103adef76e8f69a632977"} Dec 04 19:13:55 crc kubenswrapper[4733]: I1204 19:13:55.405105 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-mgml8" podStartSLOduration=3.405080624 podStartE2EDuration="3.405080624s" podCreationTimestamp="2025-12-04 19:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:13:55.40199412 +0000 UTC m=+5697.357355186" watchObservedRunningTime="2025-12-04 19:13:55.405080624 +0000 UTC m=+5697.360441670" Dec 04 19:13:58 crc kubenswrapper[4733]: I1204 19:13:58.415528 4733 generic.go:334] "Generic (PLEG): container finished" podID="0d88cdb3-e89d-4743-a918-4485083caaf5" containerID="a60c6dc6ddfd19c65af9c8a9b2d5c196b947bfe4a4e103adef76e8f69a632977" exitCode=0 Dec 04 19:13:58 crc kubenswrapper[4733]: I1204 19:13:58.415899 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mgml8" event={"ID":"0d88cdb3-e89d-4743-a918-4485083caaf5","Type":"ContainerDied","Data":"a60c6dc6ddfd19c65af9c8a9b2d5c196b947bfe4a4e103adef76e8f69a632977"} Dec 04 19:13:59 crc kubenswrapper[4733]: I1204 19:13:59.821724 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mgml8" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:13:59.999844 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-config-data\") pod \"0d88cdb3-e89d-4743-a918-4485083caaf5\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:13:59.999921 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-combined-ca-bundle\") pod \"0d88cdb3-e89d-4743-a918-4485083caaf5\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.000808 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms4xz\" (UniqueName: \"kubernetes.io/projected/0d88cdb3-e89d-4743-a918-4485083caaf5-kube-api-access-ms4xz\") pod \"0d88cdb3-e89d-4743-a918-4485083caaf5\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.001100 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-db-sync-config-data\") pod \"0d88cdb3-e89d-4743-a918-4485083caaf5\" (UID: \"0d88cdb3-e89d-4743-a918-4485083caaf5\") " Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.011124 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0d88cdb3-e89d-4743-a918-4485083caaf5" (UID: "0d88cdb3-e89d-4743-a918-4485083caaf5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.011242 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d88cdb3-e89d-4743-a918-4485083caaf5-kube-api-access-ms4xz" (OuterVolumeSpecName: "kube-api-access-ms4xz") pod "0d88cdb3-e89d-4743-a918-4485083caaf5" (UID: "0d88cdb3-e89d-4743-a918-4485083caaf5"). InnerVolumeSpecName "kube-api-access-ms4xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.051113 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d88cdb3-e89d-4743-a918-4485083caaf5" (UID: "0d88cdb3-e89d-4743-a918-4485083caaf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.077870 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-config-data" (OuterVolumeSpecName: "config-data") pod "0d88cdb3-e89d-4743-a918-4485083caaf5" (UID: "0d88cdb3-e89d-4743-a918-4485083caaf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.102169 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.102364 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.102428 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms4xz\" (UniqueName: \"kubernetes.io/projected/0d88cdb3-e89d-4743-a918-4485083caaf5-kube-api-access-ms4xz\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.102492 4733 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0d88cdb3-e89d-4743-a918-4485083caaf5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.439510 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mgml8" event={"ID":"0d88cdb3-e89d-4743-a918-4485083caaf5","Type":"ContainerDied","Data":"928ec252c6e281f72fba35ffb9e591070b6949cb0a4ee4059f76fc4c927ac23e"} Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.439956 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="928ec252c6e281f72fba35ffb9e591070b6949cb0a4ee4059f76fc4c927ac23e" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.439595 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mgml8" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.803236 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:00 crc kubenswrapper[4733]: E1204 19:14:00.803591 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d88cdb3-e89d-4743-a918-4485083caaf5" containerName="glance-db-sync" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.803615 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d88cdb3-e89d-4743-a918-4485083caaf5" containerName="glance-db-sync" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.803856 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d88cdb3-e89d-4743-a918-4485083caaf5" containerName="glance-db-sync" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.804821 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.806130 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.808433 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.808884 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-r76xm" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.809469 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.827853 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.899621 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869bdcb899-j4xkn"] Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.900985 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.924138 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869bdcb899-j4xkn"] Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934192 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-config\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934236 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-nb\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934293 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzrkq\" (UniqueName: \"kubernetes.io/projected/123534bf-486e-44db-8d34-c0ee5f8950e3-kube-api-access-dzrkq\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934317 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-ceph\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934452 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk4l2\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-kube-api-access-dk4l2\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934506 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934531 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-logs\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934581 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934617 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-dns-svc\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934718 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934769 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:00 crc kubenswrapper[4733]: I1204 19:14:00.934929 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-sb\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.035461 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.035513 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-dns-svc\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.035553 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.035575 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.035604 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-sb\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.035636 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-config\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.035651 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-nb\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.035688 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzrkq\" (UniqueName: \"kubernetes.io/projected/123534bf-486e-44db-8d34-c0ee5f8950e3-kube-api-access-dzrkq\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036353 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036669 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-nb\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036688 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-dns-svc\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036732 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-config\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-ceph\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036826 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk4l2\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-kube-api-access-dk4l2\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036844 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036861 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-logs\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.037127 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-logs\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.036765 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-sb\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.047562 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-ceph\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.047891 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.048449 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.057817 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.061720 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.063277 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.067227 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.071502 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk4l2\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-kube-api-access-dk4l2\") pod \"glance-default-external-api-0\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.079779 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.092842 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzrkq\" (UniqueName: \"kubernetes.io/projected/123534bf-486e-44db-8d34-c0ee5f8950e3-kube-api-access-dzrkq\") pod \"dnsmasq-dns-869bdcb899-j4xkn\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.122570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.137457 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.137547 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.137621 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.137643 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.137661 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k64f\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-kube-api-access-9k64f\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.137685 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.137721 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.227912 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.238746 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.238867 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.238933 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.238971 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.238997 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k64f\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-kube-api-access-9k64f\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.239035 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.239088 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.239511 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.240225 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.243055 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.243735 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.248530 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.248832 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.258576 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k64f\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-kube-api-access-9k64f\") pod \"glance-default-internal-api-0\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.432507 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.709477 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:01 crc kubenswrapper[4733]: W1204 19:14:01.716329 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f95793c_813d_4452_ac88_1c5d6512669d.slice/crio-01d151d942144dbe17f1eca7b17c83ec9196524862cf49b7b690b156f90418ea WatchSource:0}: Error finding container 01d151d942144dbe17f1eca7b17c83ec9196524862cf49b7b690b156f90418ea: Status 404 returned error can't find the container with id 01d151d942144dbe17f1eca7b17c83ec9196524862cf49b7b690b156f90418ea Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.743477 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869bdcb899-j4xkn"] Dec 04 19:14:01 crc kubenswrapper[4733]: W1204 19:14:01.747722 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod123534bf_486e_44db_8d34_c0ee5f8950e3.slice/crio-052e85e3832d4a75156c5ed5cacb346641ee6080c27fe13c455d2507d146f410 WatchSource:0}: Error finding container 052e85e3832d4a75156c5ed5cacb346641ee6080c27fe13c455d2507d146f410: Status 404 returned error can't find the container with id 052e85e3832d4a75156c5ed5cacb346641ee6080c27fe13c455d2507d146f410 Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.895291 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:01 crc kubenswrapper[4733]: I1204 19:14:01.988904 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:02 crc kubenswrapper[4733]: W1204 19:14:02.002012 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b824e36_46bb_41bd_a929_8181db945b66.slice/crio-e033751136caaffd11db9df8d4082e8bf36b838c920e40bb12ff23c3e0902720 WatchSource:0}: Error finding container e033751136caaffd11db9df8d4082e8bf36b838c920e40bb12ff23c3e0902720: Status 404 returned error can't find the container with id e033751136caaffd11db9df8d4082e8bf36b838c920e40bb12ff23c3e0902720 Dec 04 19:14:02 crc kubenswrapper[4733]: I1204 19:14:02.455698 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b824e36-46bb-41bd-a929-8181db945b66","Type":"ContainerStarted","Data":"e033751136caaffd11db9df8d4082e8bf36b838c920e40bb12ff23c3e0902720"} Dec 04 19:14:02 crc kubenswrapper[4733]: I1204 19:14:02.458461 4733 generic.go:334] "Generic (PLEG): container finished" podID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerID="c0d09c73c203cd843036e0cef5bfba0e458477bb96a316f84532336211bbaddf" exitCode=0 Dec 04 19:14:02 crc kubenswrapper[4733]: I1204 19:14:02.458546 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" event={"ID":"123534bf-486e-44db-8d34-c0ee5f8950e3","Type":"ContainerDied","Data":"c0d09c73c203cd843036e0cef5bfba0e458477bb96a316f84532336211bbaddf"} Dec 04 19:14:02 crc kubenswrapper[4733]: I1204 19:14:02.458573 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" event={"ID":"123534bf-486e-44db-8d34-c0ee5f8950e3","Type":"ContainerStarted","Data":"052e85e3832d4a75156c5ed5cacb346641ee6080c27fe13c455d2507d146f410"} Dec 04 19:14:02 crc kubenswrapper[4733]: I1204 19:14:02.466617 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f95793c-813d-4452-ac88-1c5d6512669d","Type":"ContainerStarted","Data":"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db"} Dec 04 19:14:02 crc kubenswrapper[4733]: I1204 19:14:02.466658 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f95793c-813d-4452-ac88-1c5d6512669d","Type":"ContainerStarted","Data":"01d151d942144dbe17f1eca7b17c83ec9196524862cf49b7b690b156f90418ea"} Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.477875 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b824e36-46bb-41bd-a929-8181db945b66","Type":"ContainerStarted","Data":"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933"} Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.478195 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b824e36-46bb-41bd-a929-8181db945b66","Type":"ContainerStarted","Data":"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7"} Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.485647 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" event={"ID":"123534bf-486e-44db-8d34-c0ee5f8950e3","Type":"ContainerStarted","Data":"fa0bcb049d7b60216df454765945ab1a15133c8e32ce060e2db1633d9932173c"} Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.486193 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.488423 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f95793c-813d-4452-ac88-1c5d6512669d","Type":"ContainerStarted","Data":"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d"} Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.488514 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" containerName="glance-log" containerID="cri-o://371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db" gracePeriod=30 Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.488718 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" containerName="glance-httpd" containerID="cri-o://d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d" gracePeriod=30 Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.513642 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.513597044 podStartE2EDuration="2.513597044s" podCreationTimestamp="2025-12-04 19:14:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:14:03.500102539 +0000 UTC m=+5705.455463585" watchObservedRunningTime="2025-12-04 19:14:03.513597044 +0000 UTC m=+5705.468958090" Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.525047 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" podStartSLOduration=3.525027513 podStartE2EDuration="3.525027513s" podCreationTimestamp="2025-12-04 19:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:14:03.518975418 +0000 UTC m=+5705.474336464" watchObservedRunningTime="2025-12-04 19:14:03.525027513 +0000 UTC m=+5705.480388559" Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.543195 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.5431661930000002 podStartE2EDuration="3.543166193s" podCreationTimestamp="2025-12-04 19:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:14:03.536160804 +0000 UTC m=+5705.491521850" watchObservedRunningTime="2025-12-04 19:14:03.543166193 +0000 UTC m=+5705.498527239" Dec 04 19:14:03 crc kubenswrapper[4733]: I1204 19:14:03.755852 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.111439 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.292311 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-httpd-run\") pod \"0f95793c-813d-4452-ac88-1c5d6512669d\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.292905 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0f95793c-813d-4452-ac88-1c5d6512669d" (UID: "0f95793c-813d-4452-ac88-1c5d6512669d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.293348 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-ceph\") pod \"0f95793c-813d-4452-ac88-1c5d6512669d\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.293394 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk4l2\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-kube-api-access-dk4l2\") pod \"0f95793c-813d-4452-ac88-1c5d6512669d\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.294487 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-config-data\") pod \"0f95793c-813d-4452-ac88-1c5d6512669d\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.294542 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-logs\") pod \"0f95793c-813d-4452-ac88-1c5d6512669d\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.294595 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-scripts\") pod \"0f95793c-813d-4452-ac88-1c5d6512669d\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.294649 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-combined-ca-bundle\") pod \"0f95793c-813d-4452-ac88-1c5d6512669d\" (UID: \"0f95793c-813d-4452-ac88-1c5d6512669d\") " Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.294840 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-logs" (OuterVolumeSpecName: "logs") pod "0f95793c-813d-4452-ac88-1c5d6512669d" (UID: "0f95793c-813d-4452-ac88-1c5d6512669d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.295779 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.295843 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f95793c-813d-4452-ac88-1c5d6512669d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.302568 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-scripts" (OuterVolumeSpecName: "scripts") pod "0f95793c-813d-4452-ac88-1c5d6512669d" (UID: "0f95793c-813d-4452-ac88-1c5d6512669d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.302761 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-kube-api-access-dk4l2" (OuterVolumeSpecName: "kube-api-access-dk4l2") pod "0f95793c-813d-4452-ac88-1c5d6512669d" (UID: "0f95793c-813d-4452-ac88-1c5d6512669d"). InnerVolumeSpecName "kube-api-access-dk4l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.303981 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-ceph" (OuterVolumeSpecName: "ceph") pod "0f95793c-813d-4452-ac88-1c5d6512669d" (UID: "0f95793c-813d-4452-ac88-1c5d6512669d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.327858 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f95793c-813d-4452-ac88-1c5d6512669d" (UID: "0f95793c-813d-4452-ac88-1c5d6512669d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.347728 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-config-data" (OuterVolumeSpecName: "config-data") pod "0f95793c-813d-4452-ac88-1c5d6512669d" (UID: "0f95793c-813d-4452-ac88-1c5d6512669d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.397861 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.397904 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.397921 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk4l2\" (UniqueName: \"kubernetes.io/projected/0f95793c-813d-4452-ac88-1c5d6512669d-kube-api-access-dk4l2\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.397938 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.397954 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f95793c-813d-4452-ac88-1c5d6512669d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.498753 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f95793c-813d-4452-ac88-1c5d6512669d" containerID="d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d" exitCode=0 Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.498838 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f95793c-813d-4452-ac88-1c5d6512669d" containerID="371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db" exitCode=143 Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.498946 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.499021 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f95793c-813d-4452-ac88-1c5d6512669d","Type":"ContainerDied","Data":"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d"} Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.499080 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f95793c-813d-4452-ac88-1c5d6512669d","Type":"ContainerDied","Data":"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db"} Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.499092 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f95793c-813d-4452-ac88-1c5d6512669d","Type":"ContainerDied","Data":"01d151d942144dbe17f1eca7b17c83ec9196524862cf49b7b690b156f90418ea"} Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.499107 4733 scope.go:117] "RemoveContainer" containerID="d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.525873 4733 scope.go:117] "RemoveContainer" containerID="371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.541741 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.550652 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.554777 4733 scope.go:117] "RemoveContainer" containerID="d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d" Dec 04 19:14:04 crc kubenswrapper[4733]: E1204 19:14:04.556079 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d\": container with ID starting with d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d not found: ID does not exist" containerID="d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.556122 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d"} err="failed to get container status \"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d\": rpc error: code = NotFound desc = could not find container \"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d\": container with ID starting with d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d not found: ID does not exist" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.556157 4733 scope.go:117] "RemoveContainer" containerID="371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db" Dec 04 19:14:04 crc kubenswrapper[4733]: E1204 19:14:04.556611 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db\": container with ID starting with 371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db not found: ID does not exist" containerID="371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.556654 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db"} err="failed to get container status \"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db\": rpc error: code = NotFound desc = could not find container \"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db\": container with ID starting with 371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db not found: ID does not exist" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.556681 4733 scope.go:117] "RemoveContainer" containerID="d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.557019 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d"} err="failed to get container status \"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d\": rpc error: code = NotFound desc = could not find container \"d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d\": container with ID starting with d5d052dec5da663244ad40f7798e52d1eeba10ea63c70a3dec4a4afd8bf6ab0d not found: ID does not exist" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.557052 4733 scope.go:117] "RemoveContainer" containerID="371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.557591 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db"} err="failed to get container status \"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db\": rpc error: code = NotFound desc = could not find container \"371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db\": container with ID starting with 371d2fbcd098d0efc7856180ba5178a3afa51b519aecbaefa2c073a783e525db not found: ID does not exist" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.572905 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:04 crc kubenswrapper[4733]: E1204 19:14:04.573387 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" containerName="glance-log" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.573412 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" containerName="glance-log" Dec 04 19:14:04 crc kubenswrapper[4733]: E1204 19:14:04.573448 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" containerName="glance-httpd" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.573457 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" containerName="glance-httpd" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.573690 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" containerName="glance-log" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.573719 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" containerName="glance-httpd" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.574886 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.578566 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.595338 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.608951 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-ceph\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.609007 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.609038 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-config-data\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.609084 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.609206 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-scripts\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.609287 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvgdj\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-kube-api-access-lvgdj\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.609310 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-logs\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.710868 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-ceph\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.710909 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.710929 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-config-data\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.710957 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.711004 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-scripts\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.711041 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvgdj\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-kube-api-access-lvgdj\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.711058 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-logs\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.711455 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-logs\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.713612 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.716953 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-scripts\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.717617 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.720108 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-config-data\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.721809 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-ceph\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.727590 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvgdj\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-kube-api-access-lvgdj\") pod \"glance-default-external-api-0\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " pod="openstack/glance-default-external-api-0" Dec 04 19:14:04 crc kubenswrapper[4733]: I1204 19:14:04.898555 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:14:05 crc kubenswrapper[4733]: I1204 19:14:05.477506 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:14:05 crc kubenswrapper[4733]: I1204 19:14:05.512293 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdf21970-8b52-4c07-81aa-d786baa2bb58","Type":"ContainerStarted","Data":"916df4214a2bf157cb3b95fcbd454ee09d463de0d5706a212eb8cf6a6ed01406"} Dec 04 19:14:05 crc kubenswrapper[4733]: I1204 19:14:05.512432 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7b824e36-46bb-41bd-a929-8181db945b66" containerName="glance-log" containerID="cri-o://6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7" gracePeriod=30 Dec 04 19:14:05 crc kubenswrapper[4733]: I1204 19:14:05.512939 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7b824e36-46bb-41bd-a929-8181db945b66" containerName="glance-httpd" containerID="cri-o://34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933" gracePeriod=30 Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.035162 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.142632 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-scripts\") pod \"7b824e36-46bb-41bd-a929-8181db945b66\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.142689 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k64f\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-kube-api-access-9k64f\") pod \"7b824e36-46bb-41bd-a929-8181db945b66\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.142736 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-combined-ca-bundle\") pod \"7b824e36-46bb-41bd-a929-8181db945b66\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.142768 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-httpd-run\") pod \"7b824e36-46bb-41bd-a929-8181db945b66\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.142859 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-logs\") pod \"7b824e36-46bb-41bd-a929-8181db945b66\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.143171 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7b824e36-46bb-41bd-a929-8181db945b66" (UID: "7b824e36-46bb-41bd-a929-8181db945b66"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.143353 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-logs" (OuterVolumeSpecName: "logs") pod "7b824e36-46bb-41bd-a929-8181db945b66" (UID: "7b824e36-46bb-41bd-a929-8181db945b66"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.143391 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-ceph\") pod \"7b824e36-46bb-41bd-a929-8181db945b66\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.143433 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-config-data\") pod \"7b824e36-46bb-41bd-a929-8181db945b66\" (UID: \"7b824e36-46bb-41bd-a929-8181db945b66\") " Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.143964 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.143987 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b824e36-46bb-41bd-a929-8181db945b66-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.148076 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-kube-api-access-9k64f" (OuterVolumeSpecName: "kube-api-access-9k64f") pod "7b824e36-46bb-41bd-a929-8181db945b66" (UID: "7b824e36-46bb-41bd-a929-8181db945b66"). InnerVolumeSpecName "kube-api-access-9k64f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.148186 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-ceph" (OuterVolumeSpecName: "ceph") pod "7b824e36-46bb-41bd-a929-8181db945b66" (UID: "7b824e36-46bb-41bd-a929-8181db945b66"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.148273 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-scripts" (OuterVolumeSpecName: "scripts") pod "7b824e36-46bb-41bd-a929-8181db945b66" (UID: "7b824e36-46bb-41bd-a929-8181db945b66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.182529 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b824e36-46bb-41bd-a929-8181db945b66" (UID: "7b824e36-46bb-41bd-a929-8181db945b66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.217637 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-config-data" (OuterVolumeSpecName: "config-data") pod "7b824e36-46bb-41bd-a929-8181db945b66" (UID: "7b824e36-46bb-41bd-a929-8181db945b66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.245493 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.245525 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k64f\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-kube-api-access-9k64f\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.245537 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.245547 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7b824e36-46bb-41bd-a929-8181db945b66-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.245555 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b824e36-46bb-41bd-a929-8181db945b66-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.350114 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f95793c-813d-4452-ac88-1c5d6512669d" path="/var/lib/kubelet/pods/0f95793c-813d-4452-ac88-1c5d6512669d/volumes" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.525313 4733 generic.go:334] "Generic (PLEG): container finished" podID="7b824e36-46bb-41bd-a929-8181db945b66" containerID="34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933" exitCode=0 Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.525607 4733 generic.go:334] "Generic (PLEG): container finished" podID="7b824e36-46bb-41bd-a929-8181db945b66" containerID="6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7" exitCode=143 Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.525360 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.525381 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b824e36-46bb-41bd-a929-8181db945b66","Type":"ContainerDied","Data":"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933"} Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.526553 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b824e36-46bb-41bd-a929-8181db945b66","Type":"ContainerDied","Data":"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7"} Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.526573 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b824e36-46bb-41bd-a929-8181db945b66","Type":"ContainerDied","Data":"e033751136caaffd11db9df8d4082e8bf36b838c920e40bb12ff23c3e0902720"} Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.526592 4733 scope.go:117] "RemoveContainer" containerID="34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.532373 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdf21970-8b52-4c07-81aa-d786baa2bb58","Type":"ContainerStarted","Data":"de21d2bb957d01d50a27342f3b90c33caf498e96ba91ca2639f26e7c0dccc6c4"} Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.560610 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.574770 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.586915 4733 scope.go:117] "RemoveContainer" containerID="6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.599860 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:06 crc kubenswrapper[4733]: E1204 19:14:06.600327 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b824e36-46bb-41bd-a929-8181db945b66" containerName="glance-log" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.600351 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b824e36-46bb-41bd-a929-8181db945b66" containerName="glance-log" Dec 04 19:14:06 crc kubenswrapper[4733]: E1204 19:14:06.600378 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b824e36-46bb-41bd-a929-8181db945b66" containerName="glance-httpd" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.600385 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b824e36-46bb-41bd-a929-8181db945b66" containerName="glance-httpd" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.600594 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b824e36-46bb-41bd-a929-8181db945b66" containerName="glance-log" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.600622 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b824e36-46bb-41bd-a929-8181db945b66" containerName="glance-httpd" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.601820 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.604312 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.607221 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.660384 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.660448 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.660667 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.660745 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.660774 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-logs\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.660845 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6qw8\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-kube-api-access-n6qw8\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.660874 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.703968 4733 scope.go:117] "RemoveContainer" containerID="34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933" Dec 04 19:14:06 crc kubenswrapper[4733]: E1204 19:14:06.706742 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933\": container with ID starting with 34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933 not found: ID does not exist" containerID="34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.706784 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933"} err="failed to get container status \"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933\": rpc error: code = NotFound desc = could not find container \"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933\": container with ID starting with 34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933 not found: ID does not exist" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.706823 4733 scope.go:117] "RemoveContainer" containerID="6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7" Dec 04 19:14:06 crc kubenswrapper[4733]: E1204 19:14:06.710910 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7\": container with ID starting with 6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7 not found: ID does not exist" containerID="6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.710948 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7"} err="failed to get container status \"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7\": rpc error: code = NotFound desc = could not find container \"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7\": container with ID starting with 6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7 not found: ID does not exist" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.710969 4733 scope.go:117] "RemoveContainer" containerID="34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.721958 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933"} err="failed to get container status \"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933\": rpc error: code = NotFound desc = could not find container \"34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933\": container with ID starting with 34d653acec2ccf3444b4687e8acdc9e41a334bb66ab6d88654738fc5aea63933 not found: ID does not exist" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.722002 4733 scope.go:117] "RemoveContainer" containerID="6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.725940 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7"} err="failed to get container status \"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7\": rpc error: code = NotFound desc = could not find container \"6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7\": container with ID starting with 6b442e94b939d4fa97d92b47be78a1556da9d42151f47ad8131aadf8d76d5dd7 not found: ID does not exist" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.762721 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.762785 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.762826 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.762889 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.762913 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-logs\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.762944 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6qw8\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-kube-api-access-n6qw8\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.762971 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.764047 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.769354 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-logs\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.771722 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.777129 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.777586 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.781624 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.794988 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6qw8\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-kube-api-access-n6qw8\") pod \"glance-default-internal-api-0\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:14:06 crc kubenswrapper[4733]: I1204 19:14:06.995853 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:07 crc kubenswrapper[4733]: W1204 19:14:07.540227 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda09af527_a663_4dfc_89f0_3e0d3f41c048.slice/crio-e518aee5067eec712fb3c2abf782c9c90583250da14e8d05b867fea3fc7de119 WatchSource:0}: Error finding container e518aee5067eec712fb3c2abf782c9c90583250da14e8d05b867fea3fc7de119: Status 404 returned error can't find the container with id e518aee5067eec712fb3c2abf782c9c90583250da14e8d05b867fea3fc7de119 Dec 04 19:14:07 crc kubenswrapper[4733]: I1204 19:14:07.542892 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:14:07 crc kubenswrapper[4733]: I1204 19:14:07.544502 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdf21970-8b52-4c07-81aa-d786baa2bb58","Type":"ContainerStarted","Data":"1e0372feada67ec46db547f1eb5f49dd406f57ccd058b6aefa078af1a4c96559"} Dec 04 19:14:07 crc kubenswrapper[4733]: I1204 19:14:07.580207 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.580186971 podStartE2EDuration="3.580186971s" podCreationTimestamp="2025-12-04 19:14:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:14:07.562160704 +0000 UTC m=+5709.517521770" watchObservedRunningTime="2025-12-04 19:14:07.580186971 +0000 UTC m=+5709.535548017" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.351654 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b824e36-46bb-41bd-a929-8181db945b66" path="/var/lib/kubelet/pods/7b824e36-46bb-41bd-a929-8181db945b66/volumes" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.553743 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a09af527-a663-4dfc-89f0-3e0d3f41c048","Type":"ContainerStarted","Data":"3998038ce1e96df93e35f5935b8e0f2681e565a155bc670d2d3a88053dbec451"} Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.553804 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a09af527-a663-4dfc-89f0-3e0d3f41c048","Type":"ContainerStarted","Data":"e518aee5067eec712fb3c2abf782c9c90583250da14e8d05b867fea3fc7de119"} Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.626899 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dwqjm"] Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.628951 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.642769 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwqjm"] Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.706182 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-utilities\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.706271 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmvm7\" (UniqueName: \"kubernetes.io/projected/99d26eb8-4f71-4a88-954c-ad684fa5f70d-kube-api-access-qmvm7\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.706342 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-catalog-content\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.807500 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-catalog-content\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.807942 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-utilities\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.808018 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmvm7\" (UniqueName: \"kubernetes.io/projected/99d26eb8-4f71-4a88-954c-ad684fa5f70d-kube-api-access-qmvm7\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.808043 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-catalog-content\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.808286 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-utilities\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.827728 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmvm7\" (UniqueName: \"kubernetes.io/projected/99d26eb8-4f71-4a88-954c-ad684fa5f70d-kube-api-access-qmvm7\") pod \"redhat-marketplace-dwqjm\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:08 crc kubenswrapper[4733]: I1204 19:14:08.961849 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:09 crc kubenswrapper[4733]: I1204 19:14:09.441553 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwqjm"] Dec 04 19:14:09 crc kubenswrapper[4733]: I1204 19:14:09.567453 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a09af527-a663-4dfc-89f0-3e0d3f41c048","Type":"ContainerStarted","Data":"f31aa29f94996c0ece09b0fa003c32cb17412c5b29b02e1ac58ae7b31cf6a428"} Dec 04 19:14:09 crc kubenswrapper[4733]: I1204 19:14:09.569116 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwqjm" event={"ID":"99d26eb8-4f71-4a88-954c-ad684fa5f70d","Type":"ContainerStarted","Data":"0649361c6b3d6bffe8323160cea53fd1c9a14fdf0b6fb5b805f2411b2b16d699"} Dec 04 19:14:09 crc kubenswrapper[4733]: I1204 19:14:09.601882 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.601856692 podStartE2EDuration="3.601856692s" podCreationTimestamp="2025-12-04 19:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:14:09.600303089 +0000 UTC m=+5711.555664165" watchObservedRunningTime="2025-12-04 19:14:09.601856692 +0000 UTC m=+5711.557217758" Dec 04 19:14:10 crc kubenswrapper[4733]: I1204 19:14:10.589502 4733 generic.go:334] "Generic (PLEG): container finished" podID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerID="882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2" exitCode=0 Dec 04 19:14:10 crc kubenswrapper[4733]: I1204 19:14:10.590184 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwqjm" event={"ID":"99d26eb8-4f71-4a88-954c-ad684fa5f70d","Type":"ContainerDied","Data":"882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2"} Dec 04 19:14:10 crc kubenswrapper[4733]: I1204 19:14:10.595985 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.228969 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.289537 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d466c8f4c-vn5n2"] Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.289848 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" podUID="566dedb2-20cf-44a9-97f0-47f19706d36a" containerName="dnsmasq-dns" containerID="cri-o://25f1247a2aecf9686eff30239023b84a3152d22e84c36bafed5de8505c33ce34" gracePeriod=10 Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.605316 4733 generic.go:334] "Generic (PLEG): container finished" podID="566dedb2-20cf-44a9-97f0-47f19706d36a" containerID="25f1247a2aecf9686eff30239023b84a3152d22e84c36bafed5de8505c33ce34" exitCode=0 Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.605355 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" event={"ID":"566dedb2-20cf-44a9-97f0-47f19706d36a","Type":"ContainerDied","Data":"25f1247a2aecf9686eff30239023b84a3152d22e84c36bafed5de8505c33ce34"} Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.765276 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.870830 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-dns-svc\") pod \"566dedb2-20cf-44a9-97f0-47f19706d36a\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.870875 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-nb\") pod \"566dedb2-20cf-44a9-97f0-47f19706d36a\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.870901 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-sb\") pod \"566dedb2-20cf-44a9-97f0-47f19706d36a\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.870950 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-config\") pod \"566dedb2-20cf-44a9-97f0-47f19706d36a\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.871045 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fx49v\" (UniqueName: \"kubernetes.io/projected/566dedb2-20cf-44a9-97f0-47f19706d36a-kube-api-access-fx49v\") pod \"566dedb2-20cf-44a9-97f0-47f19706d36a\" (UID: \"566dedb2-20cf-44a9-97f0-47f19706d36a\") " Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.876377 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/566dedb2-20cf-44a9-97f0-47f19706d36a-kube-api-access-fx49v" (OuterVolumeSpecName: "kube-api-access-fx49v") pod "566dedb2-20cf-44a9-97f0-47f19706d36a" (UID: "566dedb2-20cf-44a9-97f0-47f19706d36a"). InnerVolumeSpecName "kube-api-access-fx49v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.916722 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "566dedb2-20cf-44a9-97f0-47f19706d36a" (UID: "566dedb2-20cf-44a9-97f0-47f19706d36a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.923284 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-config" (OuterVolumeSpecName: "config") pod "566dedb2-20cf-44a9-97f0-47f19706d36a" (UID: "566dedb2-20cf-44a9-97f0-47f19706d36a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.926412 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "566dedb2-20cf-44a9-97f0-47f19706d36a" (UID: "566dedb2-20cf-44a9-97f0-47f19706d36a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.929262 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "566dedb2-20cf-44a9-97f0-47f19706d36a" (UID: "566dedb2-20cf-44a9-97f0-47f19706d36a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.973261 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fx49v\" (UniqueName: \"kubernetes.io/projected/566dedb2-20cf-44a9-97f0-47f19706d36a-kube-api-access-fx49v\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.973318 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.973337 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.973360 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:11 crc kubenswrapper[4733]: I1204 19:14:11.973382 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/566dedb2-20cf-44a9-97f0-47f19706d36a-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:12 crc kubenswrapper[4733]: I1204 19:14:12.617497 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" event={"ID":"566dedb2-20cf-44a9-97f0-47f19706d36a","Type":"ContainerDied","Data":"eca9d77885a1d27190003ac1b9c4c02fbc0c1429927487326b550f7abc5b08c2"} Dec 04 19:14:12 crc kubenswrapper[4733]: I1204 19:14:12.617580 4733 scope.go:117] "RemoveContainer" containerID="25f1247a2aecf9686eff30239023b84a3152d22e84c36bafed5de8505c33ce34" Dec 04 19:14:12 crc kubenswrapper[4733]: I1204 19:14:12.617578 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d466c8f4c-vn5n2" Dec 04 19:14:12 crc kubenswrapper[4733]: I1204 19:14:12.621323 4733 generic.go:334] "Generic (PLEG): container finished" podID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerID="57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0" exitCode=0 Dec 04 19:14:12 crc kubenswrapper[4733]: I1204 19:14:12.621375 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwqjm" event={"ID":"99d26eb8-4f71-4a88-954c-ad684fa5f70d","Type":"ContainerDied","Data":"57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0"} Dec 04 19:14:12 crc kubenswrapper[4733]: I1204 19:14:12.651993 4733 scope.go:117] "RemoveContainer" containerID="851d082c84fcb82e2e712fc926f773f7c91249c843d15956f59f583ff1f4d392" Dec 04 19:14:12 crc kubenswrapper[4733]: I1204 19:14:12.688380 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d466c8f4c-vn5n2"] Dec 04 19:14:12 crc kubenswrapper[4733]: I1204 19:14:12.709410 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d466c8f4c-vn5n2"] Dec 04 19:14:13 crc kubenswrapper[4733]: I1204 19:14:13.631935 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwqjm" event={"ID":"99d26eb8-4f71-4a88-954c-ad684fa5f70d","Type":"ContainerStarted","Data":"e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b"} Dec 04 19:14:14 crc kubenswrapper[4733]: I1204 19:14:14.345992 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="566dedb2-20cf-44a9-97f0-47f19706d36a" path="/var/lib/kubelet/pods/566dedb2-20cf-44a9-97f0-47f19706d36a/volumes" Dec 04 19:14:14 crc kubenswrapper[4733]: I1204 19:14:14.898827 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 19:14:14 crc kubenswrapper[4733]: I1204 19:14:14.899733 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 19:14:14 crc kubenswrapper[4733]: I1204 19:14:14.944051 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 19:14:14 crc kubenswrapper[4733]: I1204 19:14:14.979735 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dwqjm" podStartSLOduration=4.526940316 podStartE2EDuration="6.979720806s" podCreationTimestamp="2025-12-04 19:14:08 +0000 UTC" firstStartedPulling="2025-12-04 19:14:10.595527257 +0000 UTC m=+5712.550888353" lastFinishedPulling="2025-12-04 19:14:13.048307797 +0000 UTC m=+5715.003668843" observedRunningTime="2025-12-04 19:14:13.673842555 +0000 UTC m=+5715.629203621" watchObservedRunningTime="2025-12-04 19:14:14.979720806 +0000 UTC m=+5716.935081852" Dec 04 19:14:14 crc kubenswrapper[4733]: I1204 19:14:14.986144 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 19:14:15 crc kubenswrapper[4733]: I1204 19:14:15.362466 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:14:15 crc kubenswrapper[4733]: I1204 19:14:15.362529 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:14:15 crc kubenswrapper[4733]: I1204 19:14:15.657643 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 19:14:15 crc kubenswrapper[4733]: I1204 19:14:15.657688 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 19:14:16 crc kubenswrapper[4733]: I1204 19:14:16.996529 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:16 crc kubenswrapper[4733]: I1204 19:14:16.996928 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:17 crc kubenswrapper[4733]: I1204 19:14:17.034176 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:17 crc kubenswrapper[4733]: I1204 19:14:17.056322 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:17 crc kubenswrapper[4733]: I1204 19:14:17.627358 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 19:14:17 crc kubenswrapper[4733]: I1204 19:14:17.630933 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 19:14:17 crc kubenswrapper[4733]: I1204 19:14:17.681236 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:17 crc kubenswrapper[4733]: I1204 19:14:17.681293 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:18 crc kubenswrapper[4733]: I1204 19:14:18.962739 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:18 crc kubenswrapper[4733]: I1204 19:14:18.963116 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:19 crc kubenswrapper[4733]: I1204 19:14:19.025323 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:19 crc kubenswrapper[4733]: I1204 19:14:19.641060 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:19 crc kubenswrapper[4733]: I1204 19:14:19.702218 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 19:14:19 crc kubenswrapper[4733]: I1204 19:14:19.721330 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 19:14:19 crc kubenswrapper[4733]: I1204 19:14:19.779647 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:19 crc kubenswrapper[4733]: I1204 19:14:19.847134 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwqjm"] Dec 04 19:14:21 crc kubenswrapper[4733]: I1204 19:14:21.724427 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dwqjm" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerName="registry-server" containerID="cri-o://e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b" gracePeriod=2 Dec 04 19:14:21 crc kubenswrapper[4733]: I1204 19:14:21.754237 4733 scope.go:117] "RemoveContainer" containerID="a780193699f74ba0880403e8a212a23ba5e0a63948555b48c044896dc9b5e7a7" Dec 04 19:14:21 crc kubenswrapper[4733]: I1204 19:14:21.816971 4733 scope.go:117] "RemoveContainer" containerID="6fe2c2d7f492eeeceb56fe6f49ed43167571c3108bb54338838590bf5c59c7bb" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.198451 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.275056 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmvm7\" (UniqueName: \"kubernetes.io/projected/99d26eb8-4f71-4a88-954c-ad684fa5f70d-kube-api-access-qmvm7\") pod \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.275517 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-catalog-content\") pod \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.275704 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-utilities\") pod \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\" (UID: \"99d26eb8-4f71-4a88-954c-ad684fa5f70d\") " Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.278096 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-utilities" (OuterVolumeSpecName: "utilities") pod "99d26eb8-4f71-4a88-954c-ad684fa5f70d" (UID: "99d26eb8-4f71-4a88-954c-ad684fa5f70d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.283983 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d26eb8-4f71-4a88-954c-ad684fa5f70d-kube-api-access-qmvm7" (OuterVolumeSpecName: "kube-api-access-qmvm7") pod "99d26eb8-4f71-4a88-954c-ad684fa5f70d" (UID: "99d26eb8-4f71-4a88-954c-ad684fa5f70d"). InnerVolumeSpecName "kube-api-access-qmvm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.305782 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99d26eb8-4f71-4a88-954c-ad684fa5f70d" (UID: "99d26eb8-4f71-4a88-954c-ad684fa5f70d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.379917 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.379961 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmvm7\" (UniqueName: \"kubernetes.io/projected/99d26eb8-4f71-4a88-954c-ad684fa5f70d-kube-api-access-qmvm7\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.379974 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99d26eb8-4f71-4a88-954c-ad684fa5f70d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.735687 4733 generic.go:334] "Generic (PLEG): container finished" podID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerID="e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b" exitCode=0 Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.735761 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwqjm" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.735769 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwqjm" event={"ID":"99d26eb8-4f71-4a88-954c-ad684fa5f70d","Type":"ContainerDied","Data":"e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b"} Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.736549 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwqjm" event={"ID":"99d26eb8-4f71-4a88-954c-ad684fa5f70d","Type":"ContainerDied","Data":"0649361c6b3d6bffe8323160cea53fd1c9a14fdf0b6fb5b805f2411b2b16d699"} Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.736575 4733 scope.go:117] "RemoveContainer" containerID="e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.778636 4733 scope.go:117] "RemoveContainer" containerID="57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.786613 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwqjm"] Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.792486 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwqjm"] Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.802997 4733 scope.go:117] "RemoveContainer" containerID="882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.861532 4733 scope.go:117] "RemoveContainer" containerID="e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b" Dec 04 19:14:22 crc kubenswrapper[4733]: E1204 19:14:22.862122 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b\": container with ID starting with e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b not found: ID does not exist" containerID="e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.862248 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b"} err="failed to get container status \"e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b\": rpc error: code = NotFound desc = could not find container \"e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b\": container with ID starting with e3d3778bce6fccf1b5e23e099b6fa215e23c20b415162afa581b3618ae42f74b not found: ID does not exist" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.862332 4733 scope.go:117] "RemoveContainer" containerID="57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0" Dec 04 19:14:22 crc kubenswrapper[4733]: E1204 19:14:22.862604 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0\": container with ID starting with 57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0 not found: ID does not exist" containerID="57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.862681 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0"} err="failed to get container status \"57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0\": rpc error: code = NotFound desc = could not find container \"57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0\": container with ID starting with 57cdbffb91e4920274bfbc52ca64e3450b8e5e5437902545fadd470cd6058de0 not found: ID does not exist" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.862760 4733 scope.go:117] "RemoveContainer" containerID="882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2" Dec 04 19:14:22 crc kubenswrapper[4733]: E1204 19:14:22.863094 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2\": container with ID starting with 882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2 not found: ID does not exist" containerID="882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2" Dec 04 19:14:22 crc kubenswrapper[4733]: I1204 19:14:22.863413 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2"} err="failed to get container status \"882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2\": rpc error: code = NotFound desc = could not find container \"882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2\": container with ID starting with 882adbe04ab967260dd342e1d11c5cecab2ee7da2c72fd13607cc1fe272501f2 not found: ID does not exist" Dec 04 19:14:24 crc kubenswrapper[4733]: I1204 19:14:24.349438 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" path="/var/lib/kubelet/pods/99d26eb8-4f71-4a88-954c-ad684fa5f70d/volumes" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.373377 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-8wgfz"] Dec 04 19:14:29 crc kubenswrapper[4733]: E1204 19:14:29.374363 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerName="extract-utilities" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.374382 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerName="extract-utilities" Dec 04 19:14:29 crc kubenswrapper[4733]: E1204 19:14:29.374406 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566dedb2-20cf-44a9-97f0-47f19706d36a" containerName="init" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.374414 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="566dedb2-20cf-44a9-97f0-47f19706d36a" containerName="init" Dec 04 19:14:29 crc kubenswrapper[4733]: E1204 19:14:29.374427 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerName="registry-server" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.374435 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerName="registry-server" Dec 04 19:14:29 crc kubenswrapper[4733]: E1204 19:14:29.374471 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerName="extract-content" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.374480 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerName="extract-content" Dec 04 19:14:29 crc kubenswrapper[4733]: E1204 19:14:29.374492 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566dedb2-20cf-44a9-97f0-47f19706d36a" containerName="dnsmasq-dns" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.374499 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="566dedb2-20cf-44a9-97f0-47f19706d36a" containerName="dnsmasq-dns" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.374702 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="566dedb2-20cf-44a9-97f0-47f19706d36a" containerName="dnsmasq-dns" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.374726 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d26eb8-4f71-4a88-954c-ad684fa5f70d" containerName="registry-server" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.375443 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.382099 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8wgfz"] Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.475180 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-fd5b-account-create-update-pmmtr"] Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.476252 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.477956 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.490976 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fd5b-account-create-update-pmmtr"] Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.509243 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zws6c\" (UniqueName: \"kubernetes.io/projected/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-kube-api-access-zws6c\") pod \"placement-db-create-8wgfz\" (UID: \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\") " pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.509453 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-operator-scripts\") pod \"placement-db-create-8wgfz\" (UID: \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\") " pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.611150 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-operator-scripts\") pod \"placement-db-create-8wgfz\" (UID: \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\") " pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.611217 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9643222a-15e6-40a3-b978-68e777c387d5-operator-scripts\") pod \"placement-fd5b-account-create-update-pmmtr\" (UID: \"9643222a-15e6-40a3-b978-68e777c387d5\") " pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.611267 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zws6c\" (UniqueName: \"kubernetes.io/projected/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-kube-api-access-zws6c\") pod \"placement-db-create-8wgfz\" (UID: \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\") " pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.611308 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4dm9\" (UniqueName: \"kubernetes.io/projected/9643222a-15e6-40a3-b978-68e777c387d5-kube-api-access-q4dm9\") pod \"placement-fd5b-account-create-update-pmmtr\" (UID: \"9643222a-15e6-40a3-b978-68e777c387d5\") " pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.611924 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-operator-scripts\") pod \"placement-db-create-8wgfz\" (UID: \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\") " pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.633027 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zws6c\" (UniqueName: \"kubernetes.io/projected/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-kube-api-access-zws6c\") pod \"placement-db-create-8wgfz\" (UID: \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\") " pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.703295 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.713470 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9643222a-15e6-40a3-b978-68e777c387d5-operator-scripts\") pod \"placement-fd5b-account-create-update-pmmtr\" (UID: \"9643222a-15e6-40a3-b978-68e777c387d5\") " pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.713576 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4dm9\" (UniqueName: \"kubernetes.io/projected/9643222a-15e6-40a3-b978-68e777c387d5-kube-api-access-q4dm9\") pod \"placement-fd5b-account-create-update-pmmtr\" (UID: \"9643222a-15e6-40a3-b978-68e777c387d5\") " pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.715121 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9643222a-15e6-40a3-b978-68e777c387d5-operator-scripts\") pod \"placement-fd5b-account-create-update-pmmtr\" (UID: \"9643222a-15e6-40a3-b978-68e777c387d5\") " pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.733205 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4dm9\" (UniqueName: \"kubernetes.io/projected/9643222a-15e6-40a3-b978-68e777c387d5-kube-api-access-q4dm9\") pod \"placement-fd5b-account-create-update-pmmtr\" (UID: \"9643222a-15e6-40a3-b978-68e777c387d5\") " pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:29 crc kubenswrapper[4733]: I1204 19:14:29.795851 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:30 crc kubenswrapper[4733]: I1204 19:14:30.203355 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8wgfz"] Dec 04 19:14:30 crc kubenswrapper[4733]: I1204 19:14:30.307024 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fd5b-account-create-update-pmmtr"] Dec 04 19:14:30 crc kubenswrapper[4733]: W1204 19:14:30.308293 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9643222a_15e6_40a3_b978_68e777c387d5.slice/crio-0b7af18d1db65801088eccd248b30f25301687696c69e84fbdf39cfd547927f3 WatchSource:0}: Error finding container 0b7af18d1db65801088eccd248b30f25301687696c69e84fbdf39cfd547927f3: Status 404 returned error can't find the container with id 0b7af18d1db65801088eccd248b30f25301687696c69e84fbdf39cfd547927f3 Dec 04 19:14:30 crc kubenswrapper[4733]: I1204 19:14:30.815986 4733 generic.go:334] "Generic (PLEG): container finished" podID="9643222a-15e6-40a3-b978-68e777c387d5" containerID="f6dbdc315efef701e580c60f2c1db14ec2b00a9d7b4bd2fe8bdfb0533a5091fe" exitCode=0 Dec 04 19:14:30 crc kubenswrapper[4733]: I1204 19:14:30.816072 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fd5b-account-create-update-pmmtr" event={"ID":"9643222a-15e6-40a3-b978-68e777c387d5","Type":"ContainerDied","Data":"f6dbdc315efef701e580c60f2c1db14ec2b00a9d7b4bd2fe8bdfb0533a5091fe"} Dec 04 19:14:30 crc kubenswrapper[4733]: I1204 19:14:30.816299 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fd5b-account-create-update-pmmtr" event={"ID":"9643222a-15e6-40a3-b978-68e777c387d5","Type":"ContainerStarted","Data":"0b7af18d1db65801088eccd248b30f25301687696c69e84fbdf39cfd547927f3"} Dec 04 19:14:30 crc kubenswrapper[4733]: I1204 19:14:30.819019 4733 generic.go:334] "Generic (PLEG): container finished" podID="e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1" containerID="f3459b98e6bd5c66f51153a0d7f2b5801385e9875f57698c03c3a40b15965f6d" exitCode=0 Dec 04 19:14:30 crc kubenswrapper[4733]: I1204 19:14:30.819058 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8wgfz" event={"ID":"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1","Type":"ContainerDied","Data":"f3459b98e6bd5c66f51153a0d7f2b5801385e9875f57698c03c3a40b15965f6d"} Dec 04 19:14:30 crc kubenswrapper[4733]: I1204 19:14:30.819126 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8wgfz" event={"ID":"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1","Type":"ContainerStarted","Data":"050253ee176ef8879fa44cedf2bfd1d7c7db5c015addd93b7a031684ed235a63"} Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.217343 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.227321 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.397327 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zws6c\" (UniqueName: \"kubernetes.io/projected/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-kube-api-access-zws6c\") pod \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\" (UID: \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\") " Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.397401 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9643222a-15e6-40a3-b978-68e777c387d5-operator-scripts\") pod \"9643222a-15e6-40a3-b978-68e777c387d5\" (UID: \"9643222a-15e6-40a3-b978-68e777c387d5\") " Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.397488 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-operator-scripts\") pod \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\" (UID: \"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1\") " Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.397563 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4dm9\" (UniqueName: \"kubernetes.io/projected/9643222a-15e6-40a3-b978-68e777c387d5-kube-api-access-q4dm9\") pod \"9643222a-15e6-40a3-b978-68e777c387d5\" (UID: \"9643222a-15e6-40a3-b978-68e777c387d5\") " Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.398393 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1" (UID: "e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.398393 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9643222a-15e6-40a3-b978-68e777c387d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9643222a-15e6-40a3-b978-68e777c387d5" (UID: "9643222a-15e6-40a3-b978-68e777c387d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.403232 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-kube-api-access-zws6c" (OuterVolumeSpecName: "kube-api-access-zws6c") pod "e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1" (UID: "e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1"). InnerVolumeSpecName "kube-api-access-zws6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.404185 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9643222a-15e6-40a3-b978-68e777c387d5-kube-api-access-q4dm9" (OuterVolumeSpecName: "kube-api-access-q4dm9") pod "9643222a-15e6-40a3-b978-68e777c387d5" (UID: "9643222a-15e6-40a3-b978-68e777c387d5"). InnerVolumeSpecName "kube-api-access-q4dm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.499368 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zws6c\" (UniqueName: \"kubernetes.io/projected/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-kube-api-access-zws6c\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.499964 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9643222a-15e6-40a3-b978-68e777c387d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.499981 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.499994 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4dm9\" (UniqueName: \"kubernetes.io/projected/9643222a-15e6-40a3-b978-68e777c387d5-kube-api-access-q4dm9\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.840084 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fd5b-account-create-update-pmmtr" event={"ID":"9643222a-15e6-40a3-b978-68e777c387d5","Type":"ContainerDied","Data":"0b7af18d1db65801088eccd248b30f25301687696c69e84fbdf39cfd547927f3"} Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.840105 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fd5b-account-create-update-pmmtr" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.840707 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b7af18d1db65801088eccd248b30f25301687696c69e84fbdf39cfd547927f3" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.842691 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8wgfz" event={"ID":"e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1","Type":"ContainerDied","Data":"050253ee176ef8879fa44cedf2bfd1d7c7db5c015addd93b7a031684ed235a63"} Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.842717 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="050253ee176ef8879fa44cedf2bfd1d7c7db5c015addd93b7a031684ed235a63" Dec 04 19:14:32 crc kubenswrapper[4733]: I1204 19:14:32.842766 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8wgfz" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.807566 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c98688f7-wh4vs"] Dec 04 19:14:34 crc kubenswrapper[4733]: E1204 19:14:34.808237 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9643222a-15e6-40a3-b978-68e777c387d5" containerName="mariadb-account-create-update" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.808252 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9643222a-15e6-40a3-b978-68e777c387d5" containerName="mariadb-account-create-update" Dec 04 19:14:34 crc kubenswrapper[4733]: E1204 19:14:34.808268 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1" containerName="mariadb-database-create" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.808274 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1" containerName="mariadb-database-create" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.808453 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1" containerName="mariadb-database-create" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.808473 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9643222a-15e6-40a3-b978-68e777c387d5" containerName="mariadb-account-create-update" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.809656 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.826191 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c98688f7-wh4vs"] Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.862985 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-bdqb6"] Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.868190 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.875609 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-795j4" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.879108 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.879360 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.887700 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-bdqb6"] Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.942984 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-sb\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.943070 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-config\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.943107 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x8zc\" (UniqueName: \"kubernetes.io/projected/9c881d71-9509-4330-a864-76b596c5b5a6-kube-api-access-6x8zc\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.943150 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-dns-svc\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:34 crc kubenswrapper[4733]: I1204 19:14:34.943193 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-nb\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045281 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v8kg\" (UniqueName: \"kubernetes.io/projected/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-kube-api-access-9v8kg\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045360 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-dns-svc\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045411 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-nb\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045451 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-config-data\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045488 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-logs\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045517 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-scripts\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045559 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-sb\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045589 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-combined-ca-bundle\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045623 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-config\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.045658 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x8zc\" (UniqueName: \"kubernetes.io/projected/9c881d71-9509-4330-a864-76b596c5b5a6-kube-api-access-6x8zc\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.047085 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-dns-svc\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.047730 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-nb\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.048456 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-sb\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.049025 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-config\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.082714 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x8zc\" (UniqueName: \"kubernetes.io/projected/9c881d71-9509-4330-a864-76b596c5b5a6-kube-api-access-6x8zc\") pod \"dnsmasq-dns-8c98688f7-wh4vs\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.147042 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.147902 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v8kg\" (UniqueName: \"kubernetes.io/projected/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-kube-api-access-9v8kg\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.147992 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-config-data\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.148027 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-logs\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.148052 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-scripts\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.148094 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-combined-ca-bundle\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.148515 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-logs\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.151865 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-scripts\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.152052 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-config-data\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.152260 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-combined-ca-bundle\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.169100 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v8kg\" (UniqueName: \"kubernetes.io/projected/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-kube-api-access-9v8kg\") pod \"placement-db-sync-bdqb6\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.213940 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.628417 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c98688f7-wh4vs"] Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.766840 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-bdqb6"] Dec 04 19:14:35 crc kubenswrapper[4733]: W1204 19:14:35.769084 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fe62d0c_ea33_47e8_9cf0_f81a9bbaee10.slice/crio-d8a41ea2d378ebd0e9c56482699fab0866f493afe7d0a97e29281a048a8fd863 WatchSource:0}: Error finding container d8a41ea2d378ebd0e9c56482699fab0866f493afe7d0a97e29281a048a8fd863: Status 404 returned error can't find the container with id d8a41ea2d378ebd0e9c56482699fab0866f493afe7d0a97e29281a048a8fd863 Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.868091 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" event={"ID":"9c881d71-9509-4330-a864-76b596c5b5a6","Type":"ContainerStarted","Data":"5db5c03c5b144fb6fa12d8281d8e3d2c7eef6115507f894708d36ad54fbacb44"} Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.868137 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" event={"ID":"9c881d71-9509-4330-a864-76b596c5b5a6","Type":"ContainerStarted","Data":"eb3ee9a6cb7d97683cb700937ca8395e2af8f27c323fa9c8aca5c1021b4b5d32"} Dec 04 19:14:35 crc kubenswrapper[4733]: I1204 19:14:35.868980 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bdqb6" event={"ID":"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10","Type":"ContainerStarted","Data":"d8a41ea2d378ebd0e9c56482699fab0866f493afe7d0a97e29281a048a8fd863"} Dec 04 19:14:36 crc kubenswrapper[4733]: I1204 19:14:36.880486 4733 generic.go:334] "Generic (PLEG): container finished" podID="9c881d71-9509-4330-a864-76b596c5b5a6" containerID="5db5c03c5b144fb6fa12d8281d8e3d2c7eef6115507f894708d36ad54fbacb44" exitCode=0 Dec 04 19:14:36 crc kubenswrapper[4733]: I1204 19:14:36.880593 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" event={"ID":"9c881d71-9509-4330-a864-76b596c5b5a6","Type":"ContainerDied","Data":"5db5c03c5b144fb6fa12d8281d8e3d2c7eef6115507f894708d36ad54fbacb44"} Dec 04 19:14:36 crc kubenswrapper[4733]: I1204 19:14:36.886349 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bdqb6" event={"ID":"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10","Type":"ContainerStarted","Data":"b42d4642c74b57585a1c8002ad88a3d3181f2a5e79c9190632788fb62733d585"} Dec 04 19:14:36 crc kubenswrapper[4733]: I1204 19:14:36.930576 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-bdqb6" podStartSLOduration=2.930552445 podStartE2EDuration="2.930552445s" podCreationTimestamp="2025-12-04 19:14:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:14:36.926514606 +0000 UTC m=+5738.881875662" watchObservedRunningTime="2025-12-04 19:14:36.930552445 +0000 UTC m=+5738.885913501" Dec 04 19:14:37 crc kubenswrapper[4733]: I1204 19:14:37.898440 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" event={"ID":"9c881d71-9509-4330-a864-76b596c5b5a6","Type":"ContainerStarted","Data":"c0cdcad58a2195437cd399d9ce9d08245438bb282b9d84822f1d456e4bebad04"} Dec 04 19:14:37 crc kubenswrapper[4733]: I1204 19:14:37.898845 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:37 crc kubenswrapper[4733]: I1204 19:14:37.899903 4733 generic.go:334] "Generic (PLEG): container finished" podID="2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" containerID="b42d4642c74b57585a1c8002ad88a3d3181f2a5e79c9190632788fb62733d585" exitCode=0 Dec 04 19:14:37 crc kubenswrapper[4733]: I1204 19:14:37.899942 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bdqb6" event={"ID":"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10","Type":"ContainerDied","Data":"b42d4642c74b57585a1c8002ad88a3d3181f2a5e79c9190632788fb62733d585"} Dec 04 19:14:37 crc kubenswrapper[4733]: I1204 19:14:37.922984 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" podStartSLOduration=3.922960907 podStartE2EDuration="3.922960907s" podCreationTimestamp="2025-12-04 19:14:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:14:37.916582204 +0000 UTC m=+5739.871943270" watchObservedRunningTime="2025-12-04 19:14:37.922960907 +0000 UTC m=+5739.878321973" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.282025 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.438979 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v8kg\" (UniqueName: \"kubernetes.io/projected/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-kube-api-access-9v8kg\") pod \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.439507 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-scripts\") pod \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.439587 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-logs\") pod \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.439637 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-config-data\") pod \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.439733 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-combined-ca-bundle\") pod \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\" (UID: \"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10\") " Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.439942 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-logs" (OuterVolumeSpecName: "logs") pod "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" (UID: "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.441473 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.444754 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-kube-api-access-9v8kg" (OuterVolumeSpecName: "kube-api-access-9v8kg") pod "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" (UID: "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10"). InnerVolumeSpecName "kube-api-access-9v8kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.446923 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-scripts" (OuterVolumeSpecName: "scripts") pod "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" (UID: "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.469570 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-config-data" (OuterVolumeSpecName: "config-data") pod "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" (UID: "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.471712 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" (UID: "2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.543225 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v8kg\" (UniqueName: \"kubernetes.io/projected/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-kube-api-access-9v8kg\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.543269 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.543284 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.543325 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.925889 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bdqb6" event={"ID":"2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10","Type":"ContainerDied","Data":"d8a41ea2d378ebd0e9c56482699fab0866f493afe7d0a97e29281a048a8fd863"} Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.925925 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8a41ea2d378ebd0e9c56482699fab0866f493afe7d0a97e29281a048a8fd863" Dec 04 19:14:39 crc kubenswrapper[4733]: I1204 19:14:39.925969 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bdqb6" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.127610 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-74cddd8d8b-8hsq8"] Dec 04 19:14:40 crc kubenswrapper[4733]: E1204 19:14:40.128003 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" containerName="placement-db-sync" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.128019 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" containerName="placement-db-sync" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.128187 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" containerName="placement-db-sync" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.129123 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.134585 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.135043 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-795j4" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.135342 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.155759 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-74cddd8d8b-8hsq8"] Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.255554 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-scripts\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.255622 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-config-data\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.255661 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-combined-ca-bundle\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.255719 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrlt5\" (UniqueName: \"kubernetes.io/projected/704551cd-6b64-48c0-927c-2b3a857f07a8-kube-api-access-nrlt5\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.255757 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704551cd-6b64-48c0-927c-2b3a857f07a8-logs\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.357293 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-scripts\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.357674 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-config-data\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.357714 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-combined-ca-bundle\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.357776 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrlt5\" (UniqueName: \"kubernetes.io/projected/704551cd-6b64-48c0-927c-2b3a857f07a8-kube-api-access-nrlt5\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.357827 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704551cd-6b64-48c0-927c-2b3a857f07a8-logs\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.358378 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704551cd-6b64-48c0-927c-2b3a857f07a8-logs\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.363842 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-config-data\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.367782 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-combined-ca-bundle\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.376878 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/704551cd-6b64-48c0-927c-2b3a857f07a8-scripts\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.381877 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrlt5\" (UniqueName: \"kubernetes.io/projected/704551cd-6b64-48c0-927c-2b3a857f07a8-kube-api-access-nrlt5\") pod \"placement-74cddd8d8b-8hsq8\" (UID: \"704551cd-6b64-48c0-927c-2b3a857f07a8\") " pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:40 crc kubenswrapper[4733]: I1204 19:14:40.447319 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:41 crc kubenswrapper[4733]: I1204 19:14:41.030852 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-74cddd8d8b-8hsq8"] Dec 04 19:14:41 crc kubenswrapper[4733]: W1204 19:14:41.037619 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod704551cd_6b64_48c0_927c_2b3a857f07a8.slice/crio-a894ef17b675c5265cdf2e1ab339333e67a8db131f1970b0afdeff2d2a0d5afa WatchSource:0}: Error finding container a894ef17b675c5265cdf2e1ab339333e67a8db131f1970b0afdeff2d2a0d5afa: Status 404 returned error can't find the container with id a894ef17b675c5265cdf2e1ab339333e67a8db131f1970b0afdeff2d2a0d5afa Dec 04 19:14:41 crc kubenswrapper[4733]: I1204 19:14:41.947425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74cddd8d8b-8hsq8" event={"ID":"704551cd-6b64-48c0-927c-2b3a857f07a8","Type":"ContainerStarted","Data":"f680d2aed6b104dcef1b630646d35368cd69ae163bd7f8e5f204e5f015acc682"} Dec 04 19:14:41 crc kubenswrapper[4733]: I1204 19:14:41.947706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74cddd8d8b-8hsq8" event={"ID":"704551cd-6b64-48c0-927c-2b3a857f07a8","Type":"ContainerStarted","Data":"b7169eea7044fb48f1afb0f8961317afd38fe9c37d2cd523a403f61d58f0c8d4"} Dec 04 19:14:41 crc kubenswrapper[4733]: I1204 19:14:41.947815 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:41 crc kubenswrapper[4733]: I1204 19:14:41.947826 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74cddd8d8b-8hsq8" event={"ID":"704551cd-6b64-48c0-927c-2b3a857f07a8","Type":"ContainerStarted","Data":"a894ef17b675c5265cdf2e1ab339333e67a8db131f1970b0afdeff2d2a0d5afa"} Dec 04 19:14:41 crc kubenswrapper[4733]: I1204 19:14:41.947838 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:14:41 crc kubenswrapper[4733]: I1204 19:14:41.977052 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-74cddd8d8b-8hsq8" podStartSLOduration=1.977023156 podStartE2EDuration="1.977023156s" podCreationTimestamp="2025-12-04 19:14:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:14:41.965775102 +0000 UTC m=+5743.921136138" watchObservedRunningTime="2025-12-04 19:14:41.977023156 +0000 UTC m=+5743.932384212" Dec 04 19:14:45 crc kubenswrapper[4733]: I1204 19:14:45.149072 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:14:45 crc kubenswrapper[4733]: I1204 19:14:45.232990 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869bdcb899-j4xkn"] Dec 04 19:14:45 crc kubenswrapper[4733]: I1204 19:14:45.233669 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" podUID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerName="dnsmasq-dns" containerID="cri-o://fa0bcb049d7b60216df454765945ab1a15133c8e32ce060e2db1633d9932173c" gracePeriod=10 Dec 04 19:14:45 crc kubenswrapper[4733]: I1204 19:14:45.362260 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:14:45 crc kubenswrapper[4733]: I1204 19:14:45.362309 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:14:45 crc kubenswrapper[4733]: I1204 19:14:45.992446 4733 generic.go:334] "Generic (PLEG): container finished" podID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerID="fa0bcb049d7b60216df454765945ab1a15133c8e32ce060e2db1633d9932173c" exitCode=0 Dec 04 19:14:45 crc kubenswrapper[4733]: I1204 19:14:45.992496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" event={"ID":"123534bf-486e-44db-8d34-c0ee5f8950e3","Type":"ContainerDied","Data":"fa0bcb049d7b60216df454765945ab1a15133c8e32ce060e2db1633d9932173c"} Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.246361 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.368003 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-nb\") pod \"123534bf-486e-44db-8d34-c0ee5f8950e3\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.368154 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-config\") pod \"123534bf-486e-44db-8d34-c0ee5f8950e3\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.368291 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzrkq\" (UniqueName: \"kubernetes.io/projected/123534bf-486e-44db-8d34-c0ee5f8950e3-kube-api-access-dzrkq\") pod \"123534bf-486e-44db-8d34-c0ee5f8950e3\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.368323 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-dns-svc\") pod \"123534bf-486e-44db-8d34-c0ee5f8950e3\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.368938 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-sb\") pod \"123534bf-486e-44db-8d34-c0ee5f8950e3\" (UID: \"123534bf-486e-44db-8d34-c0ee5f8950e3\") " Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.376580 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/123534bf-486e-44db-8d34-c0ee5f8950e3-kube-api-access-dzrkq" (OuterVolumeSpecName: "kube-api-access-dzrkq") pod "123534bf-486e-44db-8d34-c0ee5f8950e3" (UID: "123534bf-486e-44db-8d34-c0ee5f8950e3"). InnerVolumeSpecName "kube-api-access-dzrkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.413979 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-config" (OuterVolumeSpecName: "config") pod "123534bf-486e-44db-8d34-c0ee5f8950e3" (UID: "123534bf-486e-44db-8d34-c0ee5f8950e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.414074 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "123534bf-486e-44db-8d34-c0ee5f8950e3" (UID: "123534bf-486e-44db-8d34-c0ee5f8950e3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.414296 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "123534bf-486e-44db-8d34-c0ee5f8950e3" (UID: "123534bf-486e-44db-8d34-c0ee5f8950e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.424542 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "123534bf-486e-44db-8d34-c0ee5f8950e3" (UID: "123534bf-486e-44db-8d34-c0ee5f8950e3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.471167 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.471199 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.471209 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzrkq\" (UniqueName: \"kubernetes.io/projected/123534bf-486e-44db-8d34-c0ee5f8950e3-kube-api-access-dzrkq\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.471218 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:46 crc kubenswrapper[4733]: I1204 19:14:46.471226 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/123534bf-486e-44db-8d34-c0ee5f8950e3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:14:47 crc kubenswrapper[4733]: I1204 19:14:47.003158 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" event={"ID":"123534bf-486e-44db-8d34-c0ee5f8950e3","Type":"ContainerDied","Data":"052e85e3832d4a75156c5ed5cacb346641ee6080c27fe13c455d2507d146f410"} Dec 04 19:14:47 crc kubenswrapper[4733]: I1204 19:14:47.003455 4733 scope.go:117] "RemoveContainer" containerID="fa0bcb049d7b60216df454765945ab1a15133c8e32ce060e2db1633d9932173c" Dec 04 19:14:47 crc kubenswrapper[4733]: I1204 19:14:47.003240 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" Dec 04 19:14:47 crc kubenswrapper[4733]: I1204 19:14:47.026222 4733 scope.go:117] "RemoveContainer" containerID="c0d09c73c203cd843036e0cef5bfba0e458477bb96a316f84532336211bbaddf" Dec 04 19:14:47 crc kubenswrapper[4733]: I1204 19:14:47.052618 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869bdcb899-j4xkn"] Dec 04 19:14:47 crc kubenswrapper[4733]: I1204 19:14:47.061342 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869bdcb899-j4xkn"] Dec 04 19:14:48 crc kubenswrapper[4733]: I1204 19:14:48.354694 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="123534bf-486e-44db-8d34-c0ee5f8950e3" path="/var/lib/kubelet/pods/123534bf-486e-44db-8d34-c0ee5f8950e3/volumes" Dec 04 19:14:51 crc kubenswrapper[4733]: I1204 19:14:51.229405 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-869bdcb899-j4xkn" podUID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.46:5353: i/o timeout" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.142424 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s"] Dec 04 19:15:00 crc kubenswrapper[4733]: E1204 19:15:00.143303 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerName="dnsmasq-dns" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.143315 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerName="dnsmasq-dns" Dec 04 19:15:00 crc kubenswrapper[4733]: E1204 19:15:00.143335 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerName="init" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.143341 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerName="init" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.143546 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="123534bf-486e-44db-8d34-c0ee5f8950e3" containerName="dnsmasq-dns" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.144121 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.146591 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.146867 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.153669 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s"] Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.226653 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-config-volume\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.226774 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsj59\" (UniqueName: \"kubernetes.io/projected/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-kube-api-access-zsj59\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.226874 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-secret-volume\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.333698 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-config-volume\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.334057 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsj59\" (UniqueName: \"kubernetes.io/projected/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-kube-api-access-zsj59\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.334164 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-secret-volume\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.334842 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-config-volume\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.347266 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-secret-volume\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.354057 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsj59\" (UniqueName: \"kubernetes.io/projected/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-kube-api-access-zsj59\") pod \"collect-profiles-29414595-v2f6s\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.469321 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:00 crc kubenswrapper[4733]: I1204 19:15:00.917997 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s"] Dec 04 19:15:00 crc kubenswrapper[4733]: W1204 19:15:00.926117 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3ac4c48_5b84_4803_a16d_c3c96effa7f5.slice/crio-3f22a2caa0597eaea4e5bc6d3100e06bb6f53e68af7ea2d7234b561e1b52ff1d WatchSource:0}: Error finding container 3f22a2caa0597eaea4e5bc6d3100e06bb6f53e68af7ea2d7234b561e1b52ff1d: Status 404 returned error can't find the container with id 3f22a2caa0597eaea4e5bc6d3100e06bb6f53e68af7ea2d7234b561e1b52ff1d Dec 04 19:15:01 crc kubenswrapper[4733]: I1204 19:15:01.135348 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" event={"ID":"f3ac4c48-5b84-4803-a16d-c3c96effa7f5","Type":"ContainerStarted","Data":"54a484384f97297c17ae390955e4d4e3a9e8d9c5381d7df03e915d2ba526887f"} Dec 04 19:15:01 crc kubenswrapper[4733]: I1204 19:15:01.135662 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" event={"ID":"f3ac4c48-5b84-4803-a16d-c3c96effa7f5","Type":"ContainerStarted","Data":"3f22a2caa0597eaea4e5bc6d3100e06bb6f53e68af7ea2d7234b561e1b52ff1d"} Dec 04 19:15:01 crc kubenswrapper[4733]: I1204 19:15:01.155172 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" podStartSLOduration=1.155154214 podStartE2EDuration="1.155154214s" podCreationTimestamp="2025-12-04 19:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:15:01.149264544 +0000 UTC m=+5763.104625611" watchObservedRunningTime="2025-12-04 19:15:01.155154214 +0000 UTC m=+5763.110515260" Dec 04 19:15:02 crc kubenswrapper[4733]: I1204 19:15:02.144391 4733 generic.go:334] "Generic (PLEG): container finished" podID="f3ac4c48-5b84-4803-a16d-c3c96effa7f5" containerID="54a484384f97297c17ae390955e4d4e3a9e8d9c5381d7df03e915d2ba526887f" exitCode=0 Dec 04 19:15:02 crc kubenswrapper[4733]: I1204 19:15:02.144443 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" event={"ID":"f3ac4c48-5b84-4803-a16d-c3c96effa7f5","Type":"ContainerDied","Data":"54a484384f97297c17ae390955e4d4e3a9e8d9c5381d7df03e915d2ba526887f"} Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.487163 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.591388 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-config-volume\") pod \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.591430 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsj59\" (UniqueName: \"kubernetes.io/projected/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-kube-api-access-zsj59\") pod \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.591539 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-secret-volume\") pod \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\" (UID: \"f3ac4c48-5b84-4803-a16d-c3c96effa7f5\") " Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.592109 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-config-volume" (OuterVolumeSpecName: "config-volume") pod "f3ac4c48-5b84-4803-a16d-c3c96effa7f5" (UID: "f3ac4c48-5b84-4803-a16d-c3c96effa7f5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.599559 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-kube-api-access-zsj59" (OuterVolumeSpecName: "kube-api-access-zsj59") pod "f3ac4c48-5b84-4803-a16d-c3c96effa7f5" (UID: "f3ac4c48-5b84-4803-a16d-c3c96effa7f5"). InnerVolumeSpecName "kube-api-access-zsj59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.601136 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f3ac4c48-5b84-4803-a16d-c3c96effa7f5" (UID: "f3ac4c48-5b84-4803-a16d-c3c96effa7f5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.693266 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.693316 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsj59\" (UniqueName: \"kubernetes.io/projected/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-kube-api-access-zsj59\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:03 crc kubenswrapper[4733]: I1204 19:15:03.693327 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f3ac4c48-5b84-4803-a16d-c3c96effa7f5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:04 crc kubenswrapper[4733]: I1204 19:15:04.166109 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" event={"ID":"f3ac4c48-5b84-4803-a16d-c3c96effa7f5","Type":"ContainerDied","Data":"3f22a2caa0597eaea4e5bc6d3100e06bb6f53e68af7ea2d7234b561e1b52ff1d"} Dec 04 19:15:04 crc kubenswrapper[4733]: I1204 19:15:04.166413 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f22a2caa0597eaea4e5bc6d3100e06bb6f53e68af7ea2d7234b561e1b52ff1d" Dec 04 19:15:04 crc kubenswrapper[4733]: I1204 19:15:04.166244 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s" Dec 04 19:15:04 crc kubenswrapper[4733]: I1204 19:15:04.219397 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4"] Dec 04 19:15:04 crc kubenswrapper[4733]: I1204 19:15:04.226705 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414550-r7lf4"] Dec 04 19:15:04 crc kubenswrapper[4733]: I1204 19:15:04.362487 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9d57197-10b5-4af4-b43a-3626efd322c3" path="/var/lib/kubelet/pods/a9d57197-10b5-4af4-b43a-3626efd322c3/volumes" Dec 04 19:15:11 crc kubenswrapper[4733]: I1204 19:15:11.542267 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:15:11 crc kubenswrapper[4733]: I1204 19:15:11.544016 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-74cddd8d8b-8hsq8" Dec 04 19:15:15 crc kubenswrapper[4733]: I1204 19:15:15.362155 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:15:15 crc kubenswrapper[4733]: I1204 19:15:15.362547 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:15:15 crc kubenswrapper[4733]: I1204 19:15:15.362603 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:15:15 crc kubenswrapper[4733]: I1204 19:15:15.363567 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:15:15 crc kubenswrapper[4733]: I1204 19:15:15.363639 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" gracePeriod=600 Dec 04 19:15:15 crc kubenswrapper[4733]: E1204 19:15:15.494685 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:15:16 crc kubenswrapper[4733]: I1204 19:15:16.270868 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" exitCode=0 Dec 04 19:15:16 crc kubenswrapper[4733]: I1204 19:15:16.270971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0"} Dec 04 19:15:16 crc kubenswrapper[4733]: I1204 19:15:16.271269 4733 scope.go:117] "RemoveContainer" containerID="ef4c43597b7d9b6a5f9198424285408b276a51a3896ec710afa65d9b6bbf8b2b" Dec 04 19:15:16 crc kubenswrapper[4733]: I1204 19:15:16.271894 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:15:16 crc kubenswrapper[4733]: E1204 19:15:16.272225 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:15:22 crc kubenswrapper[4733]: I1204 19:15:22.032586 4733 scope.go:117] "RemoveContainer" containerID="d2745a6816337974701f65b2cae3728835beb963bd60f93c9802c2c97dadcd64" Dec 04 19:15:31 crc kubenswrapper[4733]: I1204 19:15:31.335513 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:15:31 crc kubenswrapper[4733]: E1204 19:15:31.336436 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.301098 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-rv6rz"] Dec 04 19:15:35 crc kubenswrapper[4733]: E1204 19:15:35.301665 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ac4c48-5b84-4803-a16d-c3c96effa7f5" containerName="collect-profiles" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.301678 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ac4c48-5b84-4803-a16d-c3c96effa7f5" containerName="collect-profiles" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.301867 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ac4c48-5b84-4803-a16d-c3c96effa7f5" containerName="collect-profiles" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.302474 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.314999 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rv6rz"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.387768 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/673f8f69-9bd9-40c3-b67b-26f43f04afe7-operator-scripts\") pod \"nova-api-db-create-rv6rz\" (UID: \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\") " pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.387979 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4gvp\" (UniqueName: \"kubernetes.io/projected/673f8f69-9bd9-40c3-b67b-26f43f04afe7-kube-api-access-w4gvp\") pod \"nova-api-db-create-rv6rz\" (UID: \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\") " pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.389919 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-h29sl"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.390999 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.414275 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-633b-account-create-update-wx467"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.416471 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.419402 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.422920 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-h29sl"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.435969 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-633b-account-create-update-wx467"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.489860 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/673f8f69-9bd9-40c3-b67b-26f43f04afe7-operator-scripts\") pod \"nova-api-db-create-rv6rz\" (UID: \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\") " pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.489946 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-operator-scripts\") pod \"nova-cell0-db-create-h29sl\" (UID: \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\") " pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.490000 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4gvp\" (UniqueName: \"kubernetes.io/projected/673f8f69-9bd9-40c3-b67b-26f43f04afe7-kube-api-access-w4gvp\") pod \"nova-api-db-create-rv6rz\" (UID: \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\") " pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.490097 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gljj4\" (UniqueName: \"kubernetes.io/projected/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-kube-api-access-gljj4\") pod \"nova-cell0-db-create-h29sl\" (UID: \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\") " pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.490673 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/673f8f69-9bd9-40c3-b67b-26f43f04afe7-operator-scripts\") pod \"nova-api-db-create-rv6rz\" (UID: \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\") " pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.558533 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4gvp\" (UniqueName: \"kubernetes.io/projected/673f8f69-9bd9-40c3-b67b-26f43f04afe7-kube-api-access-w4gvp\") pod \"nova-api-db-create-rv6rz\" (UID: \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\") " pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.592673 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gljj4\" (UniqueName: \"kubernetes.io/projected/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-kube-api-access-gljj4\") pod \"nova-cell0-db-create-h29sl\" (UID: \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\") " pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.592774 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a7bee5-b6c4-4f00-8317-af0fe4133000-operator-scripts\") pod \"nova-api-633b-account-create-update-wx467\" (UID: \"97a7bee5-b6c4-4f00-8317-af0fe4133000\") " pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.592854 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scsbt\" (UniqueName: \"kubernetes.io/projected/97a7bee5-b6c4-4f00-8317-af0fe4133000-kube-api-access-scsbt\") pod \"nova-api-633b-account-create-update-wx467\" (UID: \"97a7bee5-b6c4-4f00-8317-af0fe4133000\") " pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.592892 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-operator-scripts\") pod \"nova-cell0-db-create-h29sl\" (UID: \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\") " pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.593536 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-operator-scripts\") pod \"nova-cell0-db-create-h29sl\" (UID: \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\") " pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.605783 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-hbrmg"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.606853 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.627681 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.633297 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8c5f-account-create-update-m2vkx"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.634374 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.639359 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gljj4\" (UniqueName: \"kubernetes.io/projected/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-kube-api-access-gljj4\") pod \"nova-cell0-db-create-h29sl\" (UID: \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\") " pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.644339 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.664580 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-hbrmg"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.698850 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8c5f-account-create-update-m2vkx"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.699787 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-operator-scripts\") pod \"nova-cell1-db-create-hbrmg\" (UID: \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\") " pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.699872 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a7bee5-b6c4-4f00-8317-af0fe4133000-operator-scripts\") pod \"nova-api-633b-account-create-update-wx467\" (UID: \"97a7bee5-b6c4-4f00-8317-af0fe4133000\") " pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.699899 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8fqs\" (UniqueName: \"kubernetes.io/projected/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-kube-api-access-x8fqs\") pod \"nova-cell1-db-create-hbrmg\" (UID: \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\") " pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.699943 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scsbt\" (UniqueName: \"kubernetes.io/projected/97a7bee5-b6c4-4f00-8317-af0fe4133000-kube-api-access-scsbt\") pod \"nova-api-633b-account-create-update-wx467\" (UID: \"97a7bee5-b6c4-4f00-8317-af0fe4133000\") " pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.700931 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a7bee5-b6c4-4f00-8317-af0fe4133000-operator-scripts\") pod \"nova-api-633b-account-create-update-wx467\" (UID: \"97a7bee5-b6c4-4f00-8317-af0fe4133000\") " pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.711035 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.741455 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scsbt\" (UniqueName: \"kubernetes.io/projected/97a7bee5-b6c4-4f00-8317-af0fe4133000-kube-api-access-scsbt\") pod \"nova-api-633b-account-create-update-wx467\" (UID: \"97a7bee5-b6c4-4f00-8317-af0fe4133000\") " pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.801460 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/628cfd34-aecf-4401-961d-42e8b4884a5d-operator-scripts\") pod \"nova-cell0-8c5f-account-create-update-m2vkx\" (UID: \"628cfd34-aecf-4401-961d-42e8b4884a5d\") " pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.801777 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qwxp\" (UniqueName: \"kubernetes.io/projected/628cfd34-aecf-4401-961d-42e8b4884a5d-kube-api-access-8qwxp\") pod \"nova-cell0-8c5f-account-create-update-m2vkx\" (UID: \"628cfd34-aecf-4401-961d-42e8b4884a5d\") " pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.801911 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-operator-scripts\") pod \"nova-cell1-db-create-hbrmg\" (UID: \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\") " pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.802057 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8fqs\" (UniqueName: \"kubernetes.io/projected/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-kube-api-access-x8fqs\") pod \"nova-cell1-db-create-hbrmg\" (UID: \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\") " pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.803485 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-operator-scripts\") pod \"nova-cell1-db-create-hbrmg\" (UID: \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\") " pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.814710 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-41e5-account-create-update-w82s7"] Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.815871 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.822413 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.822792 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8fqs\" (UniqueName: \"kubernetes.io/projected/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-kube-api-access-x8fqs\") pod \"nova-cell1-db-create-hbrmg\" (UID: \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\") " pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:35 crc kubenswrapper[4733]: I1204 19:15:35.840226 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-41e5-account-create-update-w82s7"] Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:35.904417 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/628cfd34-aecf-4401-961d-42e8b4884a5d-operator-scripts\") pod \"nova-cell0-8c5f-account-create-update-m2vkx\" (UID: \"628cfd34-aecf-4401-961d-42e8b4884a5d\") " pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:35.904474 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qwxp\" (UniqueName: \"kubernetes.io/projected/628cfd34-aecf-4401-961d-42e8b4884a5d-kube-api-access-8qwxp\") pod \"nova-cell0-8c5f-account-create-update-m2vkx\" (UID: \"628cfd34-aecf-4401-961d-42e8b4884a5d\") " pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:35.905185 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/628cfd34-aecf-4401-961d-42e8b4884a5d-operator-scripts\") pod \"nova-cell0-8c5f-account-create-update-m2vkx\" (UID: \"628cfd34-aecf-4401-961d-42e8b4884a5d\") " pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:35.924091 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qwxp\" (UniqueName: \"kubernetes.io/projected/628cfd34-aecf-4401-961d-42e8b4884a5d-kube-api-access-8qwxp\") pod \"nova-cell0-8c5f-account-create-update-m2vkx\" (UID: \"628cfd34-aecf-4401-961d-42e8b4884a5d\") " pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:35.933950 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.006660 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4146c6-9c30-4565-b821-3bd9c60e9917-operator-scripts\") pod \"nova-cell1-41e5-account-create-update-w82s7\" (UID: \"9a4146c6-9c30-4565-b821-3bd9c60e9917\") " pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.006802 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2p8f\" (UniqueName: \"kubernetes.io/projected/9a4146c6-9c30-4565-b821-3bd9c60e9917-kube-api-access-m2p8f\") pod \"nova-cell1-41e5-account-create-update-w82s7\" (UID: \"9a4146c6-9c30-4565-b821-3bd9c60e9917\") " pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.033309 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.090231 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.108005 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2p8f\" (UniqueName: \"kubernetes.io/projected/9a4146c6-9c30-4565-b821-3bd9c60e9917-kube-api-access-m2p8f\") pod \"nova-cell1-41e5-account-create-update-w82s7\" (UID: \"9a4146c6-9c30-4565-b821-3bd9c60e9917\") " pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.108127 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4146c6-9c30-4565-b821-3bd9c60e9917-operator-scripts\") pod \"nova-cell1-41e5-account-create-update-w82s7\" (UID: \"9a4146c6-9c30-4565-b821-3bd9c60e9917\") " pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.109303 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4146c6-9c30-4565-b821-3bd9c60e9917-operator-scripts\") pod \"nova-cell1-41e5-account-create-update-w82s7\" (UID: \"9a4146c6-9c30-4565-b821-3bd9c60e9917\") " pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.126513 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2p8f\" (UniqueName: \"kubernetes.io/projected/9a4146c6-9c30-4565-b821-3bd9c60e9917-kube-api-access-m2p8f\") pod \"nova-cell1-41e5-account-create-update-w82s7\" (UID: \"9a4146c6-9c30-4565-b821-3bd9c60e9917\") " pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.161689 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.871275 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rv6rz"] Dec 04 19:15:36 crc kubenswrapper[4733]: W1204 19:15:36.875988 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod673f8f69_9bd9_40c3_b67b_26f43f04afe7.slice/crio-60b00751152a29c44497c1331da01b1989271849b27f4e941abfbaf091bd0ebd WatchSource:0}: Error finding container 60b00751152a29c44497c1331da01b1989271849b27f4e941abfbaf091bd0ebd: Status 404 returned error can't find the container with id 60b00751152a29c44497c1331da01b1989271849b27f4e941abfbaf091bd0ebd Dec 04 19:15:36 crc kubenswrapper[4733]: W1204 19:15:36.970710 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dfb8e08_b816_4747_a6a0_a0e72ff9acec.slice/crio-12a9252d471d358ab3c2a5c82bdec640bcd5107d63e6bd268014cc8d5ec332c2 WatchSource:0}: Error finding container 12a9252d471d358ab3c2a5c82bdec640bcd5107d63e6bd268014cc8d5ec332c2: Status 404 returned error can't find the container with id 12a9252d471d358ab3c2a5c82bdec640bcd5107d63e6bd268014cc8d5ec332c2 Dec 04 19:15:36 crc kubenswrapper[4733]: I1204 19:15:36.980406 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-hbrmg"] Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.005778 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-41e5-account-create-update-w82s7"] Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.016902 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-h29sl"] Dec 04 19:15:37 crc kubenswrapper[4733]: W1204 19:15:37.021184 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a4146c6_9c30_4565_b821_3bd9c60e9917.slice/crio-165ca59199b1395c495b0711878abb1132a877ebf284212385fde4c3e1694a97 WatchSource:0}: Error finding container 165ca59199b1395c495b0711878abb1132a877ebf284212385fde4c3e1694a97: Status 404 returned error can't find the container with id 165ca59199b1395c495b0711878abb1132a877ebf284212385fde4c3e1694a97 Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.024776 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-633b-account-create-update-wx467"] Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.032159 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8c5f-account-create-update-m2vkx"] Dec 04 19:15:37 crc kubenswrapper[4733]: W1204 19:15:37.034290 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97a7bee5_b6c4_4f00_8317_af0fe4133000.slice/crio-793ce8483579ba98854a95a357bf9d053a22c95ba23031e6405846ab5e6df1f8 WatchSource:0}: Error finding container 793ce8483579ba98854a95a357bf9d053a22c95ba23031e6405846ab5e6df1f8: Status 404 returned error can't find the container with id 793ce8483579ba98854a95a357bf9d053a22c95ba23031e6405846ab5e6df1f8 Dec 04 19:15:37 crc kubenswrapper[4733]: W1204 19:15:37.035712 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94ac3286_fd11_4ffa_9da1_c35564d4e7a6.slice/crio-e7d24c74c0af0bc55519abb6740d299934fecab460d8ad6cd948c11d85d60167 WatchSource:0}: Error finding container e7d24c74c0af0bc55519abb6740d299934fecab460d8ad6cd948c11d85d60167: Status 404 returned error can't find the container with id e7d24c74c0af0bc55519abb6740d299934fecab460d8ad6cd948c11d85d60167 Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.477945 4733 generic.go:334] "Generic (PLEG): container finished" podID="8dfb8e08-b816-4747-a6a0-a0e72ff9acec" containerID="e0332ba56c48fe20015bc1c0ab7072b5413abed8bb199432361d772645e5bbf4" exitCode=0 Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.478082 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hbrmg" event={"ID":"8dfb8e08-b816-4747-a6a0-a0e72ff9acec","Type":"ContainerDied","Data":"e0332ba56c48fe20015bc1c0ab7072b5413abed8bb199432361d772645e5bbf4"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.478286 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hbrmg" event={"ID":"8dfb8e08-b816-4747-a6a0-a0e72ff9acec","Type":"ContainerStarted","Data":"12a9252d471d358ab3c2a5c82bdec640bcd5107d63e6bd268014cc8d5ec332c2"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.481039 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-633b-account-create-update-wx467" event={"ID":"97a7bee5-b6c4-4f00-8317-af0fe4133000","Type":"ContainerStarted","Data":"d8204971181bedeff30bcd0c95edeff010107aac8d52df9b92c4cbcb18c54b1e"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.481070 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-633b-account-create-update-wx467" event={"ID":"97a7bee5-b6c4-4f00-8317-af0fe4133000","Type":"ContainerStarted","Data":"793ce8483579ba98854a95a357bf9d053a22c95ba23031e6405846ab5e6df1f8"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.484334 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" event={"ID":"628cfd34-aecf-4401-961d-42e8b4884a5d","Type":"ContainerStarted","Data":"7cd08c488d0c6ca5def3ec8d745da047167f4633748fbd19c5d90476d7d7a815"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.484382 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" event={"ID":"628cfd34-aecf-4401-961d-42e8b4884a5d","Type":"ContainerStarted","Data":"0c78c842cfc4006fa1c60e8fa4b54843db3ead1ea7d3776502079913677f95bd"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.491314 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rv6rz" event={"ID":"673f8f69-9bd9-40c3-b67b-26f43f04afe7","Type":"ContainerStarted","Data":"ec18881b2092dca99136a1c7fcb83ae0cadac1d0d6aa515c5abd7c66428824fa"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.491400 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rv6rz" event={"ID":"673f8f69-9bd9-40c3-b67b-26f43f04afe7","Type":"ContainerStarted","Data":"60b00751152a29c44497c1331da01b1989271849b27f4e941abfbaf091bd0ebd"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.494601 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h29sl" event={"ID":"94ac3286-fd11-4ffa-9da1-c35564d4e7a6","Type":"ContainerStarted","Data":"1872a59391720f11d26825f0507162694ab176452ece560cbe421b193d86366e"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.494648 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h29sl" event={"ID":"94ac3286-fd11-4ffa-9da1-c35564d4e7a6","Type":"ContainerStarted","Data":"e7d24c74c0af0bc55519abb6740d299934fecab460d8ad6cd948c11d85d60167"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.500994 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" event={"ID":"9a4146c6-9c30-4565-b821-3bd9c60e9917","Type":"ContainerStarted","Data":"7d3370fc0d1062dbde2b3e7e7b8b8545f725b4e2e0e812d5213d7020756f61b9"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.501032 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" event={"ID":"9a4146c6-9c30-4565-b821-3bd9c60e9917","Type":"ContainerStarted","Data":"165ca59199b1395c495b0711878abb1132a877ebf284212385fde4c3e1694a97"} Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.518913 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-h29sl" podStartSLOduration=2.518895647 podStartE2EDuration="2.518895647s" podCreationTimestamp="2025-12-04 19:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:15:37.516688457 +0000 UTC m=+5799.472049513" watchObservedRunningTime="2025-12-04 19:15:37.518895647 +0000 UTC m=+5799.474256693" Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.536464 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-rv6rz" podStartSLOduration=2.536446022 podStartE2EDuration="2.536446022s" podCreationTimestamp="2025-12-04 19:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:15:37.53527954 +0000 UTC m=+5799.490640586" watchObservedRunningTime="2025-12-04 19:15:37.536446022 +0000 UTC m=+5799.491807068" Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.555259 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-633b-account-create-update-wx467" podStartSLOduration=2.55524242 podStartE2EDuration="2.55524242s" podCreationTimestamp="2025-12-04 19:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:15:37.550362548 +0000 UTC m=+5799.505723594" watchObservedRunningTime="2025-12-04 19:15:37.55524242 +0000 UTC m=+5799.510603456" Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.576876 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" podStartSLOduration=2.576849515 podStartE2EDuration="2.576849515s" podCreationTimestamp="2025-12-04 19:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:15:37.565100587 +0000 UTC m=+5799.520461653" watchObservedRunningTime="2025-12-04 19:15:37.576849515 +0000 UTC m=+5799.532210571" Dec 04 19:15:37 crc kubenswrapper[4733]: I1204 19:15:37.593123 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" podStartSLOduration=2.593105434 podStartE2EDuration="2.593105434s" podCreationTimestamp="2025-12-04 19:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:15:37.577734768 +0000 UTC m=+5799.533095814" watchObservedRunningTime="2025-12-04 19:15:37.593105434 +0000 UTC m=+5799.548466480" Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.509495 4733 generic.go:334] "Generic (PLEG): container finished" podID="94ac3286-fd11-4ffa-9da1-c35564d4e7a6" containerID="1872a59391720f11d26825f0507162694ab176452ece560cbe421b193d86366e" exitCode=0 Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.509578 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h29sl" event={"ID":"94ac3286-fd11-4ffa-9da1-c35564d4e7a6","Type":"ContainerDied","Data":"1872a59391720f11d26825f0507162694ab176452ece560cbe421b193d86366e"} Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.511517 4733 generic.go:334] "Generic (PLEG): container finished" podID="9a4146c6-9c30-4565-b821-3bd9c60e9917" containerID="7d3370fc0d1062dbde2b3e7e7b8b8545f725b4e2e0e812d5213d7020756f61b9" exitCode=0 Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.511617 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" event={"ID":"9a4146c6-9c30-4565-b821-3bd9c60e9917","Type":"ContainerDied","Data":"7d3370fc0d1062dbde2b3e7e7b8b8545f725b4e2e0e812d5213d7020756f61b9"} Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.513502 4733 generic.go:334] "Generic (PLEG): container finished" podID="97a7bee5-b6c4-4f00-8317-af0fe4133000" containerID="d8204971181bedeff30bcd0c95edeff010107aac8d52df9b92c4cbcb18c54b1e" exitCode=0 Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.513573 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-633b-account-create-update-wx467" event={"ID":"97a7bee5-b6c4-4f00-8317-af0fe4133000","Type":"ContainerDied","Data":"d8204971181bedeff30bcd0c95edeff010107aac8d52df9b92c4cbcb18c54b1e"} Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.514860 4733 generic.go:334] "Generic (PLEG): container finished" podID="628cfd34-aecf-4401-961d-42e8b4884a5d" containerID="7cd08c488d0c6ca5def3ec8d745da047167f4633748fbd19c5d90476d7d7a815" exitCode=0 Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.514938 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" event={"ID":"628cfd34-aecf-4401-961d-42e8b4884a5d","Type":"ContainerDied","Data":"7cd08c488d0c6ca5def3ec8d745da047167f4633748fbd19c5d90476d7d7a815"} Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.516478 4733 generic.go:334] "Generic (PLEG): container finished" podID="673f8f69-9bd9-40c3-b67b-26f43f04afe7" containerID="ec18881b2092dca99136a1c7fcb83ae0cadac1d0d6aa515c5abd7c66428824fa" exitCode=0 Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.516541 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rv6rz" event={"ID":"673f8f69-9bd9-40c3-b67b-26f43f04afe7","Type":"ContainerDied","Data":"ec18881b2092dca99136a1c7fcb83ae0cadac1d0d6aa515c5abd7c66428824fa"} Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.838463 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.962853 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-operator-scripts\") pod \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\" (UID: \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\") " Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.962941 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8fqs\" (UniqueName: \"kubernetes.io/projected/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-kube-api-access-x8fqs\") pod \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\" (UID: \"8dfb8e08-b816-4747-a6a0-a0e72ff9acec\") " Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.963536 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8dfb8e08-b816-4747-a6a0-a0e72ff9acec" (UID: "8dfb8e08-b816-4747-a6a0-a0e72ff9acec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:15:38 crc kubenswrapper[4733]: I1204 19:15:38.974051 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-kube-api-access-x8fqs" (OuterVolumeSpecName: "kube-api-access-x8fqs") pod "8dfb8e08-b816-4747-a6a0-a0e72ff9acec" (UID: "8dfb8e08-b816-4747-a6a0-a0e72ff9acec"). InnerVolumeSpecName "kube-api-access-x8fqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:15:39 crc kubenswrapper[4733]: I1204 19:15:39.064584 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:39 crc kubenswrapper[4733]: I1204 19:15:39.064892 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8fqs\" (UniqueName: \"kubernetes.io/projected/8dfb8e08-b816-4747-a6a0-a0e72ff9acec-kube-api-access-x8fqs\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:39 crc kubenswrapper[4733]: I1204 19:15:39.526531 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hbrmg" event={"ID":"8dfb8e08-b816-4747-a6a0-a0e72ff9acec","Type":"ContainerDied","Data":"12a9252d471d358ab3c2a5c82bdec640bcd5107d63e6bd268014cc8d5ec332c2"} Dec 04 19:15:39 crc kubenswrapper[4733]: I1204 19:15:39.526576 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12a9252d471d358ab3c2a5c82bdec640bcd5107d63e6bd268014cc8d5ec332c2" Dec 04 19:15:39 crc kubenswrapper[4733]: I1204 19:15:39.526701 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hbrmg" Dec 04 19:15:39 crc kubenswrapper[4733]: I1204 19:15:39.912262 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.084652 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2p8f\" (UniqueName: \"kubernetes.io/projected/9a4146c6-9c30-4565-b821-3bd9c60e9917-kube-api-access-m2p8f\") pod \"9a4146c6-9c30-4565-b821-3bd9c60e9917\" (UID: \"9a4146c6-9c30-4565-b821-3bd9c60e9917\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.084810 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4146c6-9c30-4565-b821-3bd9c60e9917-operator-scripts\") pod \"9a4146c6-9c30-4565-b821-3bd9c60e9917\" (UID: \"9a4146c6-9c30-4565-b821-3bd9c60e9917\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.086169 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4146c6-9c30-4565-b821-3bd9c60e9917-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a4146c6-9c30-4565-b821-3bd9c60e9917" (UID: "9a4146c6-9c30-4565-b821-3bd9c60e9917"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.092028 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4146c6-9c30-4565-b821-3bd9c60e9917-kube-api-access-m2p8f" (OuterVolumeSpecName: "kube-api-access-m2p8f") pod "9a4146c6-9c30-4565-b821-3bd9c60e9917" (UID: "9a4146c6-9c30-4565-b821-3bd9c60e9917"). InnerVolumeSpecName "kube-api-access-m2p8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.170013 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.176245 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.183739 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.187001 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2p8f\" (UniqueName: \"kubernetes.io/projected/9a4146c6-9c30-4565-b821-3bd9c60e9917-kube-api-access-m2p8f\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.187039 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4146c6-9c30-4565-b821-3bd9c60e9917-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.195905 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.287658 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/673f8f69-9bd9-40c3-b67b-26f43f04afe7-operator-scripts\") pod \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\" (UID: \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.288381 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/673f8f69-9bd9-40c3-b67b-26f43f04afe7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "673f8f69-9bd9-40c3-b67b-26f43f04afe7" (UID: "673f8f69-9bd9-40c3-b67b-26f43f04afe7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.288640 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-operator-scripts\") pod \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\" (UID: \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.288751 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/628cfd34-aecf-4401-961d-42e8b4884a5d-operator-scripts\") pod \"628cfd34-aecf-4401-961d-42e8b4884a5d\" (UID: \"628cfd34-aecf-4401-961d-42e8b4884a5d\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.289389 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "94ac3286-fd11-4ffa-9da1-c35564d4e7a6" (UID: "94ac3286-fd11-4ffa-9da1-c35564d4e7a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.291120 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/628cfd34-aecf-4401-961d-42e8b4884a5d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "628cfd34-aecf-4401-961d-42e8b4884a5d" (UID: "628cfd34-aecf-4401-961d-42e8b4884a5d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.291344 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4gvp\" (UniqueName: \"kubernetes.io/projected/673f8f69-9bd9-40c3-b67b-26f43f04afe7-kube-api-access-w4gvp\") pod \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\" (UID: \"673f8f69-9bd9-40c3-b67b-26f43f04afe7\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.292138 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a7bee5-b6c4-4f00-8317-af0fe4133000-operator-scripts\") pod \"97a7bee5-b6c4-4f00-8317-af0fe4133000\" (UID: \"97a7bee5-b6c4-4f00-8317-af0fe4133000\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.292404 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scsbt\" (UniqueName: \"kubernetes.io/projected/97a7bee5-b6c4-4f00-8317-af0fe4133000-kube-api-access-scsbt\") pod \"97a7bee5-b6c4-4f00-8317-af0fe4133000\" (UID: \"97a7bee5-b6c4-4f00-8317-af0fe4133000\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.292597 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97a7bee5-b6c4-4f00-8317-af0fe4133000-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "97a7bee5-b6c4-4f00-8317-af0fe4133000" (UID: "97a7bee5-b6c4-4f00-8317-af0fe4133000"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.292694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qwxp\" (UniqueName: \"kubernetes.io/projected/628cfd34-aecf-4401-961d-42e8b4884a5d-kube-api-access-8qwxp\") pod \"628cfd34-aecf-4401-961d-42e8b4884a5d\" (UID: \"628cfd34-aecf-4401-961d-42e8b4884a5d\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.292818 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gljj4\" (UniqueName: \"kubernetes.io/projected/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-kube-api-access-gljj4\") pod \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\" (UID: \"94ac3286-fd11-4ffa-9da1-c35564d4e7a6\") " Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.293437 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.293456 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/628cfd34-aecf-4401-961d-42e8b4884a5d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.293465 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a7bee5-b6c4-4f00-8317-af0fe4133000-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.293474 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/673f8f69-9bd9-40c3-b67b-26f43f04afe7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.295475 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673f8f69-9bd9-40c3-b67b-26f43f04afe7-kube-api-access-w4gvp" (OuterVolumeSpecName: "kube-api-access-w4gvp") pod "673f8f69-9bd9-40c3-b67b-26f43f04afe7" (UID: "673f8f69-9bd9-40c3-b67b-26f43f04afe7"). InnerVolumeSpecName "kube-api-access-w4gvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.295897 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/628cfd34-aecf-4401-961d-42e8b4884a5d-kube-api-access-8qwxp" (OuterVolumeSpecName: "kube-api-access-8qwxp") pod "628cfd34-aecf-4401-961d-42e8b4884a5d" (UID: "628cfd34-aecf-4401-961d-42e8b4884a5d"). InnerVolumeSpecName "kube-api-access-8qwxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.297497 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a7bee5-b6c4-4f00-8317-af0fe4133000-kube-api-access-scsbt" (OuterVolumeSpecName: "kube-api-access-scsbt") pod "97a7bee5-b6c4-4f00-8317-af0fe4133000" (UID: "97a7bee5-b6c4-4f00-8317-af0fe4133000"). InnerVolumeSpecName "kube-api-access-scsbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.298330 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-kube-api-access-gljj4" (OuterVolumeSpecName: "kube-api-access-gljj4") pod "94ac3286-fd11-4ffa-9da1-c35564d4e7a6" (UID: "94ac3286-fd11-4ffa-9da1-c35564d4e7a6"). InnerVolumeSpecName "kube-api-access-gljj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.396216 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4gvp\" (UniqueName: \"kubernetes.io/projected/673f8f69-9bd9-40c3-b67b-26f43f04afe7-kube-api-access-w4gvp\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.396258 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scsbt\" (UniqueName: \"kubernetes.io/projected/97a7bee5-b6c4-4f00-8317-af0fe4133000-kube-api-access-scsbt\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.396270 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qwxp\" (UniqueName: \"kubernetes.io/projected/628cfd34-aecf-4401-961d-42e8b4884a5d-kube-api-access-8qwxp\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.396286 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gljj4\" (UniqueName: \"kubernetes.io/projected/94ac3286-fd11-4ffa-9da1-c35564d4e7a6-kube-api-access-gljj4\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.538699 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-633b-account-create-update-wx467" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.538699 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-633b-account-create-update-wx467" event={"ID":"97a7bee5-b6c4-4f00-8317-af0fe4133000","Type":"ContainerDied","Data":"793ce8483579ba98854a95a357bf9d053a22c95ba23031e6405846ab5e6df1f8"} Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.538774 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="793ce8483579ba98854a95a357bf9d053a22c95ba23031e6405846ab5e6df1f8" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.540503 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" event={"ID":"628cfd34-aecf-4401-961d-42e8b4884a5d","Type":"ContainerDied","Data":"0c78c842cfc4006fa1c60e8fa4b54843db3ead1ea7d3776502079913677f95bd"} Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.540551 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c78c842cfc4006fa1c60e8fa4b54843db3ead1ea7d3776502079913677f95bd" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.540624 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8c5f-account-create-update-m2vkx" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.543018 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rv6rz" event={"ID":"673f8f69-9bd9-40c3-b67b-26f43f04afe7","Type":"ContainerDied","Data":"60b00751152a29c44497c1331da01b1989271849b27f4e941abfbaf091bd0ebd"} Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.543061 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60b00751152a29c44497c1331da01b1989271849b27f4e941abfbaf091bd0ebd" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.543125 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rv6rz" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.545703 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h29sl" event={"ID":"94ac3286-fd11-4ffa-9da1-c35564d4e7a6","Type":"ContainerDied","Data":"e7d24c74c0af0bc55519abb6740d299934fecab460d8ad6cd948c11d85d60167"} Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.545728 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7d24c74c0af0bc55519abb6740d299934fecab460d8ad6cd948c11d85d60167" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.545774 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h29sl" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.547406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" event={"ID":"9a4146c6-9c30-4565-b821-3bd9c60e9917","Type":"ContainerDied","Data":"165ca59199b1395c495b0711878abb1132a877ebf284212385fde4c3e1694a97"} Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.547447 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="165ca59199b1395c495b0711878abb1132a877ebf284212385fde4c3e1694a97" Dec 04 19:15:40 crc kubenswrapper[4733]: I1204 19:15:40.547472 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-41e5-account-create-update-w82s7" Dec 04 19:15:42 crc kubenswrapper[4733]: I1204 19:15:42.336261 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:15:42 crc kubenswrapper[4733]: E1204 19:15:42.336964 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.801152 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-shzgf"] Dec 04 19:15:45 crc kubenswrapper[4733]: E1204 19:15:45.802886 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a7bee5-b6c4-4f00-8317-af0fe4133000" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.803244 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a7bee5-b6c4-4f00-8317-af0fe4133000" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: E1204 19:15:45.803521 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dfb8e08-b816-4747-a6a0-a0e72ff9acec" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.803631 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dfb8e08-b816-4747-a6a0-a0e72ff9acec" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: E1204 19:15:45.803767 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ac3286-fd11-4ffa-9da1-c35564d4e7a6" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.803940 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ac3286-fd11-4ffa-9da1-c35564d4e7a6" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: E1204 19:15:45.804102 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673f8f69-9bd9-40c3-b67b-26f43f04afe7" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.804223 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="673f8f69-9bd9-40c3-b67b-26f43f04afe7" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: E1204 19:15:45.804382 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="628cfd34-aecf-4401-961d-42e8b4884a5d" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.804499 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="628cfd34-aecf-4401-961d-42e8b4884a5d" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: E1204 19:15:45.804632 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4146c6-9c30-4565-b821-3bd9c60e9917" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.804748 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4146c6-9c30-4565-b821-3bd9c60e9917" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.805310 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="628cfd34-aecf-4401-961d-42e8b4884a5d" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.805448 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ac3286-fd11-4ffa-9da1-c35564d4e7a6" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.805576 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4146c6-9c30-4565-b821-3bd9c60e9917" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.805682 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dfb8e08-b816-4747-a6a0-a0e72ff9acec" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.805786 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="673f8f69-9bd9-40c3-b67b-26f43f04afe7" containerName="mariadb-database-create" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.805953 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a7bee5-b6c4-4f00-8317-af0fe4133000" containerName="mariadb-account-create-update" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.807031 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.811827 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-shzgf"] Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.814819 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xqdbp" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.814985 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.815124 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.901599 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-scripts\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.901688 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thhc6\" (UniqueName: \"kubernetes.io/projected/82b804be-19c0-4315-9104-7c4edba4cb66-kube-api-access-thhc6\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.902024 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:45 crc kubenswrapper[4733]: I1204 19:15:45.902117 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-config-data\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.003914 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.003967 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-config-data\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.004019 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-scripts\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.004053 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thhc6\" (UniqueName: \"kubernetes.io/projected/82b804be-19c0-4315-9104-7c4edba4cb66-kube-api-access-thhc6\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.018383 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-scripts\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.018462 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.019157 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-config-data\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.033780 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thhc6\" (UniqueName: \"kubernetes.io/projected/82b804be-19c0-4315-9104-7c4edba4cb66-kube-api-access-thhc6\") pod \"nova-cell0-conductor-db-sync-shzgf\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.124628 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:46 crc kubenswrapper[4733]: I1204 19:15:46.622567 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-shzgf"] Dec 04 19:15:47 crc kubenswrapper[4733]: I1204 19:15:47.610698 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-shzgf" event={"ID":"82b804be-19c0-4315-9104-7c4edba4cb66","Type":"ContainerStarted","Data":"d614fbaaf878ded8429b1ba8c8a7f3c1f24d8c0853b196bf494327fb43e06c25"} Dec 04 19:15:47 crc kubenswrapper[4733]: I1204 19:15:47.610965 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-shzgf" event={"ID":"82b804be-19c0-4315-9104-7c4edba4cb66","Type":"ContainerStarted","Data":"7097325dbd6e4e597bde49ce0e3525bf634abb68c44e25a813e883158545b99c"} Dec 04 19:15:47 crc kubenswrapper[4733]: I1204 19:15:47.630568 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-shzgf" podStartSLOduration=2.630548045 podStartE2EDuration="2.630548045s" podCreationTimestamp="2025-12-04 19:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:15:47.624139411 +0000 UTC m=+5809.579500497" watchObservedRunningTime="2025-12-04 19:15:47.630548045 +0000 UTC m=+5809.585909101" Dec 04 19:15:52 crc kubenswrapper[4733]: I1204 19:15:52.672773 4733 generic.go:334] "Generic (PLEG): container finished" podID="82b804be-19c0-4315-9104-7c4edba4cb66" containerID="d614fbaaf878ded8429b1ba8c8a7f3c1f24d8c0853b196bf494327fb43e06c25" exitCode=0 Dec 04 19:15:52 crc kubenswrapper[4733]: I1204 19:15:52.672873 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-shzgf" event={"ID":"82b804be-19c0-4315-9104-7c4edba4cb66","Type":"ContainerDied","Data":"d614fbaaf878ded8429b1ba8c8a7f3c1f24d8c0853b196bf494327fb43e06c25"} Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.076666 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.150729 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-combined-ca-bundle\") pod \"82b804be-19c0-4315-9104-7c4edba4cb66\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.150915 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thhc6\" (UniqueName: \"kubernetes.io/projected/82b804be-19c0-4315-9104-7c4edba4cb66-kube-api-access-thhc6\") pod \"82b804be-19c0-4315-9104-7c4edba4cb66\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.153036 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-config-data\") pod \"82b804be-19c0-4315-9104-7c4edba4cb66\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.153202 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-scripts\") pod \"82b804be-19c0-4315-9104-7c4edba4cb66\" (UID: \"82b804be-19c0-4315-9104-7c4edba4cb66\") " Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.163718 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b804be-19c0-4315-9104-7c4edba4cb66-kube-api-access-thhc6" (OuterVolumeSpecName: "kube-api-access-thhc6") pod "82b804be-19c0-4315-9104-7c4edba4cb66" (UID: "82b804be-19c0-4315-9104-7c4edba4cb66"). InnerVolumeSpecName "kube-api-access-thhc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.164056 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-scripts" (OuterVolumeSpecName: "scripts") pod "82b804be-19c0-4315-9104-7c4edba4cb66" (UID: "82b804be-19c0-4315-9104-7c4edba4cb66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.193033 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82b804be-19c0-4315-9104-7c4edba4cb66" (UID: "82b804be-19c0-4315-9104-7c4edba4cb66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.196568 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-config-data" (OuterVolumeSpecName: "config-data") pod "82b804be-19c0-4315-9104-7c4edba4cb66" (UID: "82b804be-19c0-4315-9104-7c4edba4cb66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.255452 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.255497 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.255512 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thhc6\" (UniqueName: \"kubernetes.io/projected/82b804be-19c0-4315-9104-7c4edba4cb66-kube-api-access-thhc6\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.255525 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82b804be-19c0-4315-9104-7c4edba4cb66-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.701612 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-shzgf" event={"ID":"82b804be-19c0-4315-9104-7c4edba4cb66","Type":"ContainerDied","Data":"7097325dbd6e4e597bde49ce0e3525bf634abb68c44e25a813e883158545b99c"} Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.702043 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7097325dbd6e4e597bde49ce0e3525bf634abb68c44e25a813e883158545b99c" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.701930 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-shzgf" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.789704 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:15:54 crc kubenswrapper[4733]: E1204 19:15:54.790459 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b804be-19c0-4315-9104-7c4edba4cb66" containerName="nova-cell0-conductor-db-sync" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.790494 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b804be-19c0-4315-9104-7c4edba4cb66" containerName="nova-cell0-conductor-db-sync" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.790874 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b804be-19c0-4315-9104-7c4edba4cb66" containerName="nova-cell0-conductor-db-sync" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.791982 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.794527 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xqdbp" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.795667 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.805265 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.866879 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.866936 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.867027 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4nk7\" (UniqueName: \"kubernetes.io/projected/dac90ac1-ccec-4016-be71-26d09ca77ae9-kube-api-access-m4nk7\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.969254 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.969309 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.969435 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4nk7\" (UniqueName: \"kubernetes.io/projected/dac90ac1-ccec-4016-be71-26d09ca77ae9-kube-api-access-m4nk7\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.977592 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.978204 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:54 crc kubenswrapper[4733]: I1204 19:15:54.985424 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4nk7\" (UniqueName: \"kubernetes.io/projected/dac90ac1-ccec-4016-be71-26d09ca77ae9-kube-api-access-m4nk7\") pod \"nova-cell0-conductor-0\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:55 crc kubenswrapper[4733]: I1204 19:15:55.113852 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:55 crc kubenswrapper[4733]: I1204 19:15:55.553156 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:15:55 crc kubenswrapper[4733]: I1204 19:15:55.709982 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dac90ac1-ccec-4016-be71-26d09ca77ae9","Type":"ContainerStarted","Data":"43ae9f0268adc4487c41224be38e23d8ab7e38bd8e74b16a6a777c28d75b46e2"} Dec 04 19:15:56 crc kubenswrapper[4733]: I1204 19:15:56.727060 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dac90ac1-ccec-4016-be71-26d09ca77ae9","Type":"ContainerStarted","Data":"ce29af541c38e6ae0a8c967e0b55f541ed938943b2097cbb748168b7b3b01aaa"} Dec 04 19:15:56 crc kubenswrapper[4733]: I1204 19:15:56.727630 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 19:15:56 crc kubenswrapper[4733]: I1204 19:15:56.755446 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.755431044 podStartE2EDuration="2.755431044s" podCreationTimestamp="2025-12-04 19:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:15:56.749979916 +0000 UTC m=+5818.705340972" watchObservedRunningTime="2025-12-04 19:15:56.755431044 +0000 UTC m=+5818.710792090" Dec 04 19:15:57 crc kubenswrapper[4733]: I1204 19:15:57.335570 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:15:57 crc kubenswrapper[4733]: E1204 19:15:57.335947 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.160283 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.684659 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-5j49q"] Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.685931 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.689659 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.690287 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.697884 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5j49q"] Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.788993 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.789074 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-scripts\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.789179 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-config-data\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.789259 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mjcm\" (UniqueName: \"kubernetes.io/projected/7c44dddb-58bb-476e-91d3-5d0663897c2e-kube-api-access-8mjcm\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.790599 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.792177 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.815736 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.843016 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.891725 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-scripts\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.891869 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-config-data\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.891924 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7kbc\" (UniqueName: \"kubernetes.io/projected/0f801e56-7e6a-420a-ba97-f161c5683757-kube-api-access-j7kbc\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.891975 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-config-data\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.892028 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mjcm\" (UniqueName: \"kubernetes.io/projected/7c44dddb-58bb-476e-91d3-5d0663897c2e-kube-api-access-8mjcm\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.892075 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.892110 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.910860 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-scripts\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.911045 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.912066 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-config-data\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.914986 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mjcm\" (UniqueName: \"kubernetes.io/projected/7c44dddb-58bb-476e-91d3-5d0663897c2e-kube-api-access-8mjcm\") pod \"nova-cell0-cell-mapping-5j49q\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.917228 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.929072 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.932281 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.966178 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.987142 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.988704 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.992177 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.993496 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7kbc\" (UniqueName: \"kubernetes.io/projected/0f801e56-7e6a-420a-ba97-f161c5683757-kube-api-access-j7kbc\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.993553 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-config-data\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.993617 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.995121 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:16:00 crc kubenswrapper[4733]: I1204 19:16:00.998322 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.003764 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.006077 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.011010 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.014781 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.015285 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-config-data\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.022532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7kbc\" (UniqueName: \"kubernetes.io/projected/0f801e56-7e6a-420a-ba97-f161c5683757-kube-api-access-j7kbc\") pod \"nova-scheduler-0\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.035938 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.095947 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-config-data\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096055 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-config-data\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096085 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzvlp\" (UniqueName: \"kubernetes.io/projected/4c3b45d1-3ff3-4b07-bf87-d5a664188440-kube-api-access-qzvlp\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096122 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjkrb\" (UniqueName: \"kubernetes.io/projected/6778ade4-c095-4dd4-b95b-ef8f660da007-kube-api-access-cjkrb\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096157 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c3b45d1-3ff3-4b07-bf87-d5a664188440-logs\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096203 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096240 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096262 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096291 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6778ade4-c095-4dd4-b95b-ef8f660da007-logs\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096337 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.096358 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2w7z\" (UniqueName: \"kubernetes.io/projected/e13167b4-439a-4071-af11-a524cd4d214e-kube-api-access-m2w7z\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.110977 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f8b566c47-9l9wx"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.112613 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.130837 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.132480 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f8b566c47-9l9wx"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198265 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-dns-svc\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-config-data\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198330 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzvlp\" (UniqueName: \"kubernetes.io/projected/4c3b45d1-3ff3-4b07-bf87-d5a664188440-kube-api-access-qzvlp\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198359 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjkrb\" (UniqueName: \"kubernetes.io/projected/6778ade4-c095-4dd4-b95b-ef8f660da007-kube-api-access-cjkrb\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198382 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcqbz\" (UniqueName: \"kubernetes.io/projected/ef390936-b761-4615-a1a9-517066f07745-kube-api-access-mcqbz\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198403 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c3b45d1-3ff3-4b07-bf87-d5a664188440-logs\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198757 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-nb\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198822 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198856 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c3b45d1-3ff3-4b07-bf87-d5a664188440-logs\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198859 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-sb\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198931 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198963 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.198987 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6778ade4-c095-4dd4-b95b-ef8f660da007-logs\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.199036 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-config\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.199054 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.199368 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6778ade4-c095-4dd4-b95b-ef8f660da007-logs\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.199683 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2w7z\" (UniqueName: \"kubernetes.io/projected/e13167b4-439a-4071-af11-a524cd4d214e-kube-api-access-m2w7z\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.199711 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-config-data\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.203905 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.206707 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.212016 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.215163 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.215824 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-config-data\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.216915 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzvlp\" (UniqueName: \"kubernetes.io/projected/4c3b45d1-3ff3-4b07-bf87-d5a664188440-kube-api-access-qzvlp\") pod \"nova-metadata-0\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.217323 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2w7z\" (UniqueName: \"kubernetes.io/projected/e13167b4-439a-4071-af11-a524cd4d214e-kube-api-access-m2w7z\") pod \"nova-cell1-novncproxy-0\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.219418 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-config-data\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.220136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjkrb\" (UniqueName: \"kubernetes.io/projected/6778ade4-c095-4dd4-b95b-ef8f660da007-kube-api-access-cjkrb\") pod \"nova-api-0\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.287761 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.301170 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcqbz\" (UniqueName: \"kubernetes.io/projected/ef390936-b761-4615-a1a9-517066f07745-kube-api-access-mcqbz\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.301238 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-nb\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.301267 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-sb\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.301381 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-config\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.301483 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-dns-svc\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.302207 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-nb\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.302344 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-config\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.303028 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-sb\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.303562 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-dns-svc\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.319465 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcqbz\" (UniqueName: \"kubernetes.io/projected/ef390936-b761-4615-a1a9-517066f07745-kube-api-access-mcqbz\") pod \"dnsmasq-dns-7f8b566c47-9l9wx\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.442177 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.466423 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.476338 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.524292 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z4fwn"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.525897 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.529461 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.529644 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.548007 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z4fwn"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.582735 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.597696 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5j49q"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.610498 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-config-data\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.610560 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-scripts\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.610771 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.610917 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zxv8\" (UniqueName: \"kubernetes.io/projected/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-kube-api-access-7zxv8\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.644412 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.715700 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-scripts\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.715748 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.715779 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zxv8\" (UniqueName: \"kubernetes.io/projected/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-kube-api-access-7zxv8\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.715880 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-config-data\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.719842 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.720013 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-config-data\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.720175 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-scripts\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.739685 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zxv8\" (UniqueName: \"kubernetes.io/projected/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-kube-api-access-7zxv8\") pod \"nova-cell1-conductor-db-sync-z4fwn\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.795149 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f801e56-7e6a-420a-ba97-f161c5683757","Type":"ContainerStarted","Data":"5418f842e3d701017e9bd3f43800bb53839e6f4db339ada0e9a4b1b086a85dee"} Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.808980 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6778ade4-c095-4dd4-b95b-ef8f660da007","Type":"ContainerStarted","Data":"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7"} Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.809045 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6778ade4-c095-4dd4-b95b-ef8f660da007","Type":"ContainerStarted","Data":"3b9454cae91de8afd0f4dd064454bc7f580234c6aa6269c16fafe11c3eb33e79"} Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.813265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5j49q" event={"ID":"7c44dddb-58bb-476e-91d3-5d0663897c2e","Type":"ContainerStarted","Data":"a9fcc81a6a766b89353a2c7d8b8fc9826941f867082ca56fd207c5b30249094f"} Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.841725 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-5j49q" podStartSLOduration=1.841707642 podStartE2EDuration="1.841707642s" podCreationTimestamp="2025-12-04 19:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:01.829228884 +0000 UTC m=+5823.784589930" watchObservedRunningTime="2025-12-04 19:16:01.841707642 +0000 UTC m=+5823.797068688" Dec 04 19:16:01 crc kubenswrapper[4733]: I1204 19:16:01.881649 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.016972 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.107692 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:02 crc kubenswrapper[4733]: W1204 19:16:02.125443 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c3b45d1_3ff3_4b07_bf87_d5a664188440.slice/crio-cfc3731b4d573d79f3d80256794ba47cbd79ae61e26d5efee97f30279f53ec54 WatchSource:0}: Error finding container cfc3731b4d573d79f3d80256794ba47cbd79ae61e26d5efee97f30279f53ec54: Status 404 returned error can't find the container with id cfc3731b4d573d79f3d80256794ba47cbd79ae61e26d5efee97f30279f53ec54 Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.130547 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f8b566c47-9l9wx"] Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.179280 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z4fwn"] Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.821735 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c3b45d1-3ff3-4b07-bf87-d5a664188440","Type":"ContainerStarted","Data":"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.825552 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c3b45d1-3ff3-4b07-bf87-d5a664188440","Type":"ContainerStarted","Data":"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.826705 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c3b45d1-3ff3-4b07-bf87-d5a664188440","Type":"ContainerStarted","Data":"cfc3731b4d573d79f3d80256794ba47cbd79ae61e26d5efee97f30279f53ec54"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.827257 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6778ade4-c095-4dd4-b95b-ef8f660da007","Type":"ContainerStarted","Data":"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.829569 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" event={"ID":"3bcff476-b8d5-4ddd-91e6-25e5c7488b57","Type":"ContainerStarted","Data":"04ef518c51ed8ba09fb4f5b2501fb3d3e0cf32c1b5a07068a443f08ff14ad56c"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.829602 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" event={"ID":"3bcff476-b8d5-4ddd-91e6-25e5c7488b57","Type":"ContainerStarted","Data":"1c5e0e5c0ce9de0f20d5cba447aaad4dfc44de046309889796904554d3ee18d0"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.831350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e13167b4-439a-4071-af11-a524cd4d214e","Type":"ContainerStarted","Data":"c662db2f16b33fb17267eee2d4d3f7d0dbece41314464b19df007b01164d6f7d"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.831390 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e13167b4-439a-4071-af11-a524cd4d214e","Type":"ContainerStarted","Data":"af87d189a952302a7c4aa69ad6641164bf5176fc1d1d143bdf0ec345e97f3828"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.833561 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5j49q" event={"ID":"7c44dddb-58bb-476e-91d3-5d0663897c2e","Type":"ContainerStarted","Data":"9a5e9e20a72da4042f0daaffa415ae764aa94ca62de0c15aa94ac58caa04baea"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.835686 4733 generic.go:334] "Generic (PLEG): container finished" podID="ef390936-b761-4615-a1a9-517066f07745" containerID="88850cf6b8ced264c71663128de23318f3409d85d1059c4234ac39ea79e79932" exitCode=0 Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.835746 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" event={"ID":"ef390936-b761-4615-a1a9-517066f07745","Type":"ContainerDied","Data":"88850cf6b8ced264c71663128de23318f3409d85d1059c4234ac39ea79e79932"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.835789 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" event={"ID":"ef390936-b761-4615-a1a9-517066f07745","Type":"ContainerStarted","Data":"2af1ad118a45d6fdeffca88dd2699f8cb2d2f6b0e0379e7bb8e93d59696dcbd4"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.838558 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f801e56-7e6a-420a-ba97-f161c5683757","Type":"ContainerStarted","Data":"8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca"} Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.854337 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.854317449 podStartE2EDuration="2.854317449s" podCreationTimestamp="2025-12-04 19:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:02.840698811 +0000 UTC m=+5824.796059857" watchObservedRunningTime="2025-12-04 19:16:02.854317449 +0000 UTC m=+5824.809678495" Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.863871 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.863855048 podStartE2EDuration="2.863855048s" podCreationTimestamp="2025-12-04 19:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:02.863045305 +0000 UTC m=+5824.818406371" watchObservedRunningTime="2025-12-04 19:16:02.863855048 +0000 UTC m=+5824.819216094" Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.883841 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.883822778 podStartE2EDuration="2.883822778s" podCreationTimestamp="2025-12-04 19:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:02.877392513 +0000 UTC m=+5824.832753559" watchObservedRunningTime="2025-12-04 19:16:02.883822778 +0000 UTC m=+5824.839183814" Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.927429 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" podStartSLOduration=1.927354125 podStartE2EDuration="1.927354125s" podCreationTimestamp="2025-12-04 19:16:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:02.917065317 +0000 UTC m=+5824.872426373" watchObservedRunningTime="2025-12-04 19:16:02.927354125 +0000 UTC m=+5824.882715181" Dec 04 19:16:02 crc kubenswrapper[4733]: I1204 19:16:02.946176 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.9461569130000003 podStartE2EDuration="2.946156913s" podCreationTimestamp="2025-12-04 19:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:02.937315324 +0000 UTC m=+5824.892676380" watchObservedRunningTime="2025-12-04 19:16:02.946156913 +0000 UTC m=+5824.901517959" Dec 04 19:16:03 crc kubenswrapper[4733]: I1204 19:16:03.852016 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" event={"ID":"ef390936-b761-4615-a1a9-517066f07745","Type":"ContainerStarted","Data":"e2f457072bddbd8e832eb075ca553f01885aa55a715bbbaf7d273270a587d110"} Dec 04 19:16:03 crc kubenswrapper[4733]: I1204 19:16:03.876268 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" podStartSLOduration=2.876250829 podStartE2EDuration="2.876250829s" podCreationTimestamp="2025-12-04 19:16:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:03.866988779 +0000 UTC m=+5825.822349845" watchObservedRunningTime="2025-12-04 19:16:03.876250829 +0000 UTC m=+5825.831611865" Dec 04 19:16:04 crc kubenswrapper[4733]: I1204 19:16:04.859663 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:05 crc kubenswrapper[4733]: I1204 19:16:05.872615 4733 generic.go:334] "Generic (PLEG): container finished" podID="3bcff476-b8d5-4ddd-91e6-25e5c7488b57" containerID="04ef518c51ed8ba09fb4f5b2501fb3d3e0cf32c1b5a07068a443f08ff14ad56c" exitCode=0 Dec 04 19:16:05 crc kubenswrapper[4733]: I1204 19:16:05.872703 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" event={"ID":"3bcff476-b8d5-4ddd-91e6-25e5c7488b57","Type":"ContainerDied","Data":"04ef518c51ed8ba09fb4f5b2501fb3d3e0cf32c1b5a07068a443f08ff14ad56c"} Dec 04 19:16:06 crc kubenswrapper[4733]: I1204 19:16:06.131720 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 19:16:06 crc kubenswrapper[4733]: I1204 19:16:06.443744 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:06 crc kubenswrapper[4733]: I1204 19:16:06.467136 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 19:16:06 crc kubenswrapper[4733]: I1204 19:16:06.467557 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 19:16:06 crc kubenswrapper[4733]: I1204 19:16:06.887444 4733 generic.go:334] "Generic (PLEG): container finished" podID="7c44dddb-58bb-476e-91d3-5d0663897c2e" containerID="9a5e9e20a72da4042f0daaffa415ae764aa94ca62de0c15aa94ac58caa04baea" exitCode=0 Dec 04 19:16:06 crc kubenswrapper[4733]: I1204 19:16:06.887533 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5j49q" event={"ID":"7c44dddb-58bb-476e-91d3-5d0663897c2e","Type":"ContainerDied","Data":"9a5e9e20a72da4042f0daaffa415ae764aa94ca62de0c15aa94ac58caa04baea"} Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.327638 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.438833 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-config-data\") pod \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.438918 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-combined-ca-bundle\") pod \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.438942 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-scripts\") pod \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.438969 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zxv8\" (UniqueName: \"kubernetes.io/projected/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-kube-api-access-7zxv8\") pod \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\" (UID: \"3bcff476-b8d5-4ddd-91e6-25e5c7488b57\") " Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.444589 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-kube-api-access-7zxv8" (OuterVolumeSpecName: "kube-api-access-7zxv8") pod "3bcff476-b8d5-4ddd-91e6-25e5c7488b57" (UID: "3bcff476-b8d5-4ddd-91e6-25e5c7488b57"). InnerVolumeSpecName "kube-api-access-7zxv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.445018 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-scripts" (OuterVolumeSpecName: "scripts") pod "3bcff476-b8d5-4ddd-91e6-25e5c7488b57" (UID: "3bcff476-b8d5-4ddd-91e6-25e5c7488b57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.463819 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-config-data" (OuterVolumeSpecName: "config-data") pod "3bcff476-b8d5-4ddd-91e6-25e5c7488b57" (UID: "3bcff476-b8d5-4ddd-91e6-25e5c7488b57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.476906 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bcff476-b8d5-4ddd-91e6-25e5c7488b57" (UID: "3bcff476-b8d5-4ddd-91e6-25e5c7488b57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.541679 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.541720 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.541735 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.541749 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zxv8\" (UniqueName: \"kubernetes.io/projected/3bcff476-b8d5-4ddd-91e6-25e5c7488b57-kube-api-access-7zxv8\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.899568 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" event={"ID":"3bcff476-b8d5-4ddd-91e6-25e5c7488b57","Type":"ContainerDied","Data":"1c5e0e5c0ce9de0f20d5cba447aaad4dfc44de046309889796904554d3ee18d0"} Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.899622 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z4fwn" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.899683 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c5e0e5c0ce9de0f20d5cba447aaad4dfc44de046309889796904554d3ee18d0" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.987623 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:16:07 crc kubenswrapper[4733]: E1204 19:16:07.988597 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcff476-b8d5-4ddd-91e6-25e5c7488b57" containerName="nova-cell1-conductor-db-sync" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.988690 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcff476-b8d5-4ddd-91e6-25e5c7488b57" containerName="nova-cell1-conductor-db-sync" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.989057 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcff476-b8d5-4ddd-91e6-25e5c7488b57" containerName="nova-cell1-conductor-db-sync" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.990060 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:07 crc kubenswrapper[4733]: I1204 19:16:07.992991 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.004020 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.052006 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc7kr\" (UniqueName: \"kubernetes.io/projected/4adf1e5a-318e-4ff3-a3e3-9af810860f12-kube-api-access-mc7kr\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.052251 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.052358 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.155134 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc7kr\" (UniqueName: \"kubernetes.io/projected/4adf1e5a-318e-4ff3-a3e3-9af810860f12-kube-api-access-mc7kr\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.155784 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.157012 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.163093 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.166486 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.183093 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc7kr\" (UniqueName: \"kubernetes.io/projected/4adf1e5a-318e-4ff3-a3e3-9af810860f12-kube-api-access-mc7kr\") pod \"nova-cell1-conductor-0\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.215360 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.321397 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.360382 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-config-data\") pod \"7c44dddb-58bb-476e-91d3-5d0663897c2e\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.360444 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-scripts\") pod \"7c44dddb-58bb-476e-91d3-5d0663897c2e\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.360560 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-combined-ca-bundle\") pod \"7c44dddb-58bb-476e-91d3-5d0663897c2e\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.360594 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mjcm\" (UniqueName: \"kubernetes.io/projected/7c44dddb-58bb-476e-91d3-5d0663897c2e-kube-api-access-8mjcm\") pod \"7c44dddb-58bb-476e-91d3-5d0663897c2e\" (UID: \"7c44dddb-58bb-476e-91d3-5d0663897c2e\") " Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.363697 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-scripts" (OuterVolumeSpecName: "scripts") pod "7c44dddb-58bb-476e-91d3-5d0663897c2e" (UID: "7c44dddb-58bb-476e-91d3-5d0663897c2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.363953 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c44dddb-58bb-476e-91d3-5d0663897c2e-kube-api-access-8mjcm" (OuterVolumeSpecName: "kube-api-access-8mjcm") pod "7c44dddb-58bb-476e-91d3-5d0663897c2e" (UID: "7c44dddb-58bb-476e-91d3-5d0663897c2e"). InnerVolumeSpecName "kube-api-access-8mjcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.389032 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c44dddb-58bb-476e-91d3-5d0663897c2e" (UID: "7c44dddb-58bb-476e-91d3-5d0663897c2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.390000 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-config-data" (OuterVolumeSpecName: "config-data") pod "7c44dddb-58bb-476e-91d3-5d0663897c2e" (UID: "7c44dddb-58bb-476e-91d3-5d0663897c2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.465596 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.465873 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mjcm\" (UniqueName: \"kubernetes.io/projected/7c44dddb-58bb-476e-91d3-5d0663897c2e-kube-api-access-8mjcm\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.465884 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.465896 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c44dddb-58bb-476e-91d3-5d0663897c2e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.756209 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.909455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5j49q" event={"ID":"7c44dddb-58bb-476e-91d3-5d0663897c2e","Type":"ContainerDied","Data":"a9fcc81a6a766b89353a2c7d8b8fc9826941f867082ca56fd207c5b30249094f"} Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.909483 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5j49q" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.909508 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9fcc81a6a766b89353a2c7d8b8fc9826941f867082ca56fd207c5b30249094f" Dec 04 19:16:08 crc kubenswrapper[4733]: I1204 19:16:08.911017 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4adf1e5a-318e-4ff3-a3e3-9af810860f12","Type":"ContainerStarted","Data":"387a2e9076b4ffceeb4bd1ec1017ebf15872efc3918fe49be5454b504c74a444"} Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.088984 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.089580 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerName="nova-api-log" containerID="cri-o://7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7" gracePeriod=30 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.089665 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerName="nova-api-api" containerID="cri-o://60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921" gracePeriod=30 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.112611 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.112876 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0f801e56-7e6a-420a-ba97-f161c5683757" containerName="nova-scheduler-scheduler" containerID="cri-o://8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca" gracePeriod=30 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.126772 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.127063 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerName="nova-metadata-log" containerID="cri-o://62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6" gracePeriod=30 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.127141 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerName="nova-metadata-metadata" containerID="cri-o://0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc" gracePeriod=30 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.652542 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.689664 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.802370 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-config-data\") pod \"6778ade4-c095-4dd4-b95b-ef8f660da007\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.802443 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzvlp\" (UniqueName: \"kubernetes.io/projected/4c3b45d1-3ff3-4b07-bf87-d5a664188440-kube-api-access-qzvlp\") pod \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.802482 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6778ade4-c095-4dd4-b95b-ef8f660da007-logs\") pod \"6778ade4-c095-4dd4-b95b-ef8f660da007\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.802597 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-config-data\") pod \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.802641 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjkrb\" (UniqueName: \"kubernetes.io/projected/6778ade4-c095-4dd4-b95b-ef8f660da007-kube-api-access-cjkrb\") pod \"6778ade4-c095-4dd4-b95b-ef8f660da007\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.802691 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-combined-ca-bundle\") pod \"6778ade4-c095-4dd4-b95b-ef8f660da007\" (UID: \"6778ade4-c095-4dd4-b95b-ef8f660da007\") " Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.802728 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c3b45d1-3ff3-4b07-bf87-d5a664188440-logs\") pod \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.802754 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-combined-ca-bundle\") pod \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\" (UID: \"4c3b45d1-3ff3-4b07-bf87-d5a664188440\") " Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.803273 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6778ade4-c095-4dd4-b95b-ef8f660da007-logs" (OuterVolumeSpecName: "logs") pod "6778ade4-c095-4dd4-b95b-ef8f660da007" (UID: "6778ade4-c095-4dd4-b95b-ef8f660da007"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.804122 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c3b45d1-3ff3-4b07-bf87-d5a664188440-logs" (OuterVolumeSpecName: "logs") pod "4c3b45d1-3ff3-4b07-bf87-d5a664188440" (UID: "4c3b45d1-3ff3-4b07-bf87-d5a664188440"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.807357 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6778ade4-c095-4dd4-b95b-ef8f660da007-kube-api-access-cjkrb" (OuterVolumeSpecName: "kube-api-access-cjkrb") pod "6778ade4-c095-4dd4-b95b-ef8f660da007" (UID: "6778ade4-c095-4dd4-b95b-ef8f660da007"). InnerVolumeSpecName "kube-api-access-cjkrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.810027 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c3b45d1-3ff3-4b07-bf87-d5a664188440-kube-api-access-qzvlp" (OuterVolumeSpecName: "kube-api-access-qzvlp") pod "4c3b45d1-3ff3-4b07-bf87-d5a664188440" (UID: "4c3b45d1-3ff3-4b07-bf87-d5a664188440"). InnerVolumeSpecName "kube-api-access-qzvlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.835319 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-config-data" (OuterVolumeSpecName: "config-data") pod "4c3b45d1-3ff3-4b07-bf87-d5a664188440" (UID: "4c3b45d1-3ff3-4b07-bf87-d5a664188440"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.837371 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-config-data" (OuterVolumeSpecName: "config-data") pod "6778ade4-c095-4dd4-b95b-ef8f660da007" (UID: "6778ade4-c095-4dd4-b95b-ef8f660da007"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.848129 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6778ade4-c095-4dd4-b95b-ef8f660da007" (UID: "6778ade4-c095-4dd4-b95b-ef8f660da007"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.852832 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c3b45d1-3ff3-4b07-bf87-d5a664188440" (UID: "4c3b45d1-3ff3-4b07-bf87-d5a664188440"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.904417 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.904451 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjkrb\" (UniqueName: \"kubernetes.io/projected/6778ade4-c095-4dd4-b95b-ef8f660da007-kube-api-access-cjkrb\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.904463 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.904471 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c3b45d1-3ff3-4b07-bf87-d5a664188440-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.904479 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3b45d1-3ff3-4b07-bf87-d5a664188440-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.904488 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6778ade4-c095-4dd4-b95b-ef8f660da007-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.904495 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzvlp\" (UniqueName: \"kubernetes.io/projected/4c3b45d1-3ff3-4b07-bf87-d5a664188440-kube-api-access-qzvlp\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.904503 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6778ade4-c095-4dd4-b95b-ef8f660da007-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.923488 4733 generic.go:334] "Generic (PLEG): container finished" podID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerID="60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921" exitCode=0 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.923517 4733 generic.go:334] "Generic (PLEG): container finished" podID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerID="7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7" exitCode=143 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.923556 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6778ade4-c095-4dd4-b95b-ef8f660da007","Type":"ContainerDied","Data":"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921"} Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.923623 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6778ade4-c095-4dd4-b95b-ef8f660da007","Type":"ContainerDied","Data":"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7"} Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.923637 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6778ade4-c095-4dd4-b95b-ef8f660da007","Type":"ContainerDied","Data":"3b9454cae91de8afd0f4dd064454bc7f580234c6aa6269c16fafe11c3eb33e79"} Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.923618 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.923687 4733 scope.go:117] "RemoveContainer" containerID="60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.926437 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4adf1e5a-318e-4ff3-a3e3-9af810860f12","Type":"ContainerStarted","Data":"51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9"} Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.929149 4733 generic.go:334] "Generic (PLEG): container finished" podID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerID="0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc" exitCode=0 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.929231 4733 generic.go:334] "Generic (PLEG): container finished" podID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerID="62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6" exitCode=143 Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.929235 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.929265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c3b45d1-3ff3-4b07-bf87-d5a664188440","Type":"ContainerDied","Data":"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc"} Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.929440 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c3b45d1-3ff3-4b07-bf87-d5a664188440","Type":"ContainerDied","Data":"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6"} Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.929450 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c3b45d1-3ff3-4b07-bf87-d5a664188440","Type":"ContainerDied","Data":"cfc3731b4d573d79f3d80256794ba47cbd79ae61e26d5efee97f30279f53ec54"} Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.951046 4733 scope.go:117] "RemoveContainer" containerID="7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.969863 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.969844971 podStartE2EDuration="2.969844971s" podCreationTimestamp="2025-12-04 19:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:09.952352549 +0000 UTC m=+5831.907713605" watchObservedRunningTime="2025-12-04 19:16:09.969844971 +0000 UTC m=+5831.925206017" Dec 04 19:16:09 crc kubenswrapper[4733]: I1204 19:16:09.982662 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.001020 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.003209 4733 scope.go:117] "RemoveContainer" containerID="60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.010081 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921\": container with ID starting with 60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921 not found: ID does not exist" containerID="60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.010129 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921"} err="failed to get container status \"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921\": rpc error: code = NotFound desc = could not find container \"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921\": container with ID starting with 60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921 not found: ID does not exist" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.011852 4733 scope.go:117] "RemoveContainer" containerID="7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.014432 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.014905 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerName="nova-api-api" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.014927 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerName="nova-api-api" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.014941 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerName="nova-metadata-metadata" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.014949 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerName="nova-metadata-metadata" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.014971 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c44dddb-58bb-476e-91d3-5d0663897c2e" containerName="nova-manage" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.014981 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c44dddb-58bb-476e-91d3-5d0663897c2e" containerName="nova-manage" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.015007 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerName="nova-metadata-log" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.015016 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerName="nova-metadata-log" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.015040 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerName="nova-api-log" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.015048 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerName="nova-api-log" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.015239 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerName="nova-metadata-metadata" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.015258 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerName="nova-api-api" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.015280 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c44dddb-58bb-476e-91d3-5d0663897c2e" containerName="nova-manage" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.015299 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" containerName="nova-api-log" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.015319 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" containerName="nova-metadata-log" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.016608 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.021948 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7\": container with ID starting with 7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7 not found: ID does not exist" containerID="7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.022009 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7"} err="failed to get container status \"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7\": rpc error: code = NotFound desc = could not find container \"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7\": container with ID starting with 7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7 not found: ID does not exist" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.022040 4733 scope.go:117] "RemoveContainer" containerID="60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.026567 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.026461 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921"} err="failed to get container status \"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921\": rpc error: code = NotFound desc = could not find container \"60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921\": container with ID starting with 60e41401235bdfb04773416e08855636e32e1185badc9dc1c47187217a788921 not found: ID does not exist" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.027507 4733 scope.go:117] "RemoveContainer" containerID="7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.029306 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7"} err="failed to get container status \"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7\": rpc error: code = NotFound desc = could not find container \"7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7\": container with ID starting with 7ae1e89620217e54a8aab87b6e552daa46b7fd588f3da9af63baacda137d8bb7 not found: ID does not exist" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.029343 4733 scope.go:117] "RemoveContainer" containerID="0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.036375 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.055654 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.069174 4733 scope.go:117] "RemoveContainer" containerID="62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.072326 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.083889 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.085757 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.088436 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.096730 4733 scope.go:117] "RemoveContainer" containerID="0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.097324 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc\": container with ID starting with 0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc not found: ID does not exist" containerID="0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.097362 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc"} err="failed to get container status \"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc\": rpc error: code = NotFound desc = could not find container \"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc\": container with ID starting with 0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc not found: ID does not exist" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.097389 4733 scope.go:117] "RemoveContainer" containerID="62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.097656 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6\": container with ID starting with 62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6 not found: ID does not exist" containerID="62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.097682 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6"} err="failed to get container status \"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6\": rpc error: code = NotFound desc = could not find container \"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6\": container with ID starting with 62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6 not found: ID does not exist" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.097699 4733 scope.go:117] "RemoveContainer" containerID="0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.098106 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc"} err="failed to get container status \"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc\": rpc error: code = NotFound desc = could not find container \"0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc\": container with ID starting with 0d241340007b383f37f3d08bd5f1299d4b2aebea0af5a97e1e6abdf9d0a3dedc not found: ID does not exist" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.098131 4733 scope.go:117] "RemoveContainer" containerID="62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.098985 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6"} err="failed to get container status \"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6\": rpc error: code = NotFound desc = could not find container \"62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6\": container with ID starting with 62afd522f51c6f6008f088a711395786a35a50a9d92429a3e9e47b2b80c4fcb6 not found: ID does not exist" Dec 04 19:16:10 crc kubenswrapper[4733]: E1204 19:16:10.101602 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c3b45d1_3ff3_4b07_bf87_d5a664188440.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6778ade4_c095_4dd4_b95b_ef8f660da007.slice/crio-3b9454cae91de8afd0f4dd064454bc7f580234c6aa6269c16fafe11c3eb33e79\": RecentStats: unable to find data in memory cache]" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.102158 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.111116 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.111172 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-config-data\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.111245 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-logs\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.111293 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgtmn\" (UniqueName: \"kubernetes.io/projected/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-kube-api-access-rgtmn\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.212708 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgtmn\" (UniqueName: \"kubernetes.io/projected/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-kube-api-access-rgtmn\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.213108 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0637413e-e16b-40dc-915e-09985282c533-logs\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.213291 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-config-data\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.213385 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.213474 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.213560 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwr4n\" (UniqueName: \"kubernetes.io/projected/0637413e-e16b-40dc-915e-09985282c533-kube-api-access-fwr4n\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.213660 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-config-data\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.213760 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-logs\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.214564 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-logs\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.220770 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-config-data\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.221734 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.234632 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgtmn\" (UniqueName: \"kubernetes.io/projected/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-kube-api-access-rgtmn\") pod \"nova-api-0\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.316258 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-config-data\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.316377 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.316450 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwr4n\" (UniqueName: \"kubernetes.io/projected/0637413e-e16b-40dc-915e-09985282c533-kube-api-access-fwr4n\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.316655 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0637413e-e16b-40dc-915e-09985282c533-logs\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.317141 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0637413e-e16b-40dc-915e-09985282c533-logs\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.320934 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.321649 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-config-data\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.338669 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwr4n\" (UniqueName: \"kubernetes.io/projected/0637413e-e16b-40dc-915e-09985282c533-kube-api-access-fwr4n\") pod \"nova-metadata-0\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.348087 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c3b45d1-3ff3-4b07-bf87-d5a664188440" path="/var/lib/kubelet/pods/4c3b45d1-3ff3-4b07-bf87-d5a664188440/volumes" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.349681 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6778ade4-c095-4dd4-b95b-ef8f660da007" path="/var/lib/kubelet/pods/6778ade4-c095-4dd4-b95b-ef8f660da007/volumes" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.362040 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.408733 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:10 crc kubenswrapper[4733]: W1204 19:16:10.887914 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd647bcd8_cc11_4c59_829b_4a4f1a568d1f.slice/crio-8eb4f0fd5f99dc569104ae1186dff6157d87dc706cf59b5382e330597c66cd32 WatchSource:0}: Error finding container 8eb4f0fd5f99dc569104ae1186dff6157d87dc706cf59b5382e330597c66cd32: Status 404 returned error can't find the container with id 8eb4f0fd5f99dc569104ae1186dff6157d87dc706cf59b5382e330597c66cd32 Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.888101 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.950334 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d647bcd8-cc11-4c59-829b-4a4f1a568d1f","Type":"ContainerStarted","Data":"8eb4f0fd5f99dc569104ae1186dff6157d87dc706cf59b5382e330597c66cd32"} Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.958662 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:10 crc kubenswrapper[4733]: I1204 19:16:10.961061 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:10 crc kubenswrapper[4733]: W1204 19:16:10.965896 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0637413e_e16b_40dc_915e_09985282c533.slice/crio-a2cb75a273ba1fe89f826ce6ff4c77f88ee475813bcad4bfaf1782982f0451a7 WatchSource:0}: Error finding container a2cb75a273ba1fe89f826ce6ff4c77f88ee475813bcad4bfaf1782982f0451a7: Status 404 returned error can't find the container with id a2cb75a273ba1fe89f826ce6ff4c77f88ee475813bcad4bfaf1782982f0451a7 Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.442698 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.456680 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.478950 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.590148 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8c98688f7-wh4vs"] Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.590367 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" podUID="9c881d71-9509-4330-a864-76b596c5b5a6" containerName="dnsmasq-dns" containerID="cri-o://c0cdcad58a2195437cd399d9ce9d08245438bb282b9d84822f1d456e4bebad04" gracePeriod=10 Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.969010 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d647bcd8-cc11-4c59-829b-4a4f1a568d1f","Type":"ContainerStarted","Data":"2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a"} Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.969421 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d647bcd8-cc11-4c59-829b-4a4f1a568d1f","Type":"ContainerStarted","Data":"2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86"} Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.970485 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0637413e-e16b-40dc-915e-09985282c533","Type":"ContainerStarted","Data":"336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1"} Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.970547 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0637413e-e16b-40dc-915e-09985282c533","Type":"ContainerStarted","Data":"b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0"} Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.970558 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0637413e-e16b-40dc-915e-09985282c533","Type":"ContainerStarted","Data":"a2cb75a273ba1fe89f826ce6ff4c77f88ee475813bcad4bfaf1782982f0451a7"} Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.973507 4733 generic.go:334] "Generic (PLEG): container finished" podID="9c881d71-9509-4330-a864-76b596c5b5a6" containerID="c0cdcad58a2195437cd399d9ce9d08245438bb282b9d84822f1d456e4bebad04" exitCode=0 Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.973580 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" event={"ID":"9c881d71-9509-4330-a864-76b596c5b5a6","Type":"ContainerDied","Data":"c0cdcad58a2195437cd399d9ce9d08245438bb282b9d84822f1d456e4bebad04"} Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.981917 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:16:11 crc kubenswrapper[4733]: I1204 19:16:11.995504 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9954842790000002 podStartE2EDuration="2.995484279s" podCreationTimestamp="2025-12-04 19:16:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:11.985047087 +0000 UTC m=+5833.940408143" watchObservedRunningTime="2025-12-04 19:16:11.995484279 +0000 UTC m=+5833.950845325" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.046379 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.046361614 podStartE2EDuration="3.046361614s" podCreationTimestamp="2025-12-04 19:16:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:12.037060563 +0000 UTC m=+5833.992421619" watchObservedRunningTime="2025-12-04 19:16:12.046361614 +0000 UTC m=+5834.001722660" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.125178 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.267221 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-nb\") pod \"9c881d71-9509-4330-a864-76b596c5b5a6\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.267285 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-config\") pod \"9c881d71-9509-4330-a864-76b596c5b5a6\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.267345 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x8zc\" (UniqueName: \"kubernetes.io/projected/9c881d71-9509-4330-a864-76b596c5b5a6-kube-api-access-6x8zc\") pod \"9c881d71-9509-4330-a864-76b596c5b5a6\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.267400 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-dns-svc\") pod \"9c881d71-9509-4330-a864-76b596c5b5a6\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.267422 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-sb\") pod \"9c881d71-9509-4330-a864-76b596c5b5a6\" (UID: \"9c881d71-9509-4330-a864-76b596c5b5a6\") " Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.280590 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c881d71-9509-4330-a864-76b596c5b5a6-kube-api-access-6x8zc" (OuterVolumeSpecName: "kube-api-access-6x8zc") pod "9c881d71-9509-4330-a864-76b596c5b5a6" (UID: "9c881d71-9509-4330-a864-76b596c5b5a6"). InnerVolumeSpecName "kube-api-access-6x8zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.312054 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9c881d71-9509-4330-a864-76b596c5b5a6" (UID: "9c881d71-9509-4330-a864-76b596c5b5a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.328347 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-config" (OuterVolumeSpecName: "config") pod "9c881d71-9509-4330-a864-76b596c5b5a6" (UID: "9c881d71-9509-4330-a864-76b596c5b5a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.333284 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9c881d71-9509-4330-a864-76b596c5b5a6" (UID: "9c881d71-9509-4330-a864-76b596c5b5a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.337261 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:16:12 crc kubenswrapper[4733]: E1204 19:16:12.337880 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.342222 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9c881d71-9509-4330-a864-76b596c5b5a6" (UID: "9c881d71-9509-4330-a864-76b596c5b5a6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.369851 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.370129 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.370244 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x8zc\" (UniqueName: \"kubernetes.io/projected/9c881d71-9509-4330-a864-76b596c5b5a6-kube-api-access-6x8zc\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.370348 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.370453 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c881d71-9509-4330-a864-76b596c5b5a6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.989637 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.990261 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c98688f7-wh4vs" event={"ID":"9c881d71-9509-4330-a864-76b596c5b5a6","Type":"ContainerDied","Data":"eb3ee9a6cb7d97683cb700937ca8395e2af8f27c323fa9c8aca5c1021b4b5d32"} Dec 04 19:16:12 crc kubenswrapper[4733]: I1204 19:16:12.990525 4733 scope.go:117] "RemoveContainer" containerID="c0cdcad58a2195437cd399d9ce9d08245438bb282b9d84822f1d456e4bebad04" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.038723 4733 scope.go:117] "RemoveContainer" containerID="5db5c03c5b144fb6fa12d8281d8e3d2c7eef6115507f894708d36ad54fbacb44" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.039633 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8c98688f7-wh4vs"] Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.066028 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8c98688f7-wh4vs"] Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.356756 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.750160 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.818048 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-6v6gj"] Dec 04 19:16:13 crc kubenswrapper[4733]: E1204 19:16:13.818694 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c881d71-9509-4330-a864-76b596c5b5a6" containerName="dnsmasq-dns" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.818723 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c881d71-9509-4330-a864-76b596c5b5a6" containerName="dnsmasq-dns" Dec 04 19:16:13 crc kubenswrapper[4733]: E1204 19:16:13.818741 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f801e56-7e6a-420a-ba97-f161c5683757" containerName="nova-scheduler-scheduler" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.818753 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f801e56-7e6a-420a-ba97-f161c5683757" containerName="nova-scheduler-scheduler" Dec 04 19:16:13 crc kubenswrapper[4733]: E1204 19:16:13.818776 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c881d71-9509-4330-a864-76b596c5b5a6" containerName="init" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.818788 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c881d71-9509-4330-a864-76b596c5b5a6" containerName="init" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.819199 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c881d71-9509-4330-a864-76b596c5b5a6" containerName="dnsmasq-dns" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.819247 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f801e56-7e6a-420a-ba97-f161c5683757" containerName="nova-scheduler-scheduler" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.820232 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.822704 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.822968 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.835149 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6v6gj"] Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.899384 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-config-data\") pod \"0f801e56-7e6a-420a-ba97-f161c5683757\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.899675 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7kbc\" (UniqueName: \"kubernetes.io/projected/0f801e56-7e6a-420a-ba97-f161c5683757-kube-api-access-j7kbc\") pod \"0f801e56-7e6a-420a-ba97-f161c5683757\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.899817 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-combined-ca-bundle\") pod \"0f801e56-7e6a-420a-ba97-f161c5683757\" (UID: \"0f801e56-7e6a-420a-ba97-f161c5683757\") " Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.900249 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.900283 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-scripts\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.900317 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-config-data\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.900374 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6vmg\" (UniqueName: \"kubernetes.io/projected/b9d10c32-2f7a-4306-82cc-f84d893f2045-kube-api-access-r6vmg\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.905190 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f801e56-7e6a-420a-ba97-f161c5683757-kube-api-access-j7kbc" (OuterVolumeSpecName: "kube-api-access-j7kbc") pod "0f801e56-7e6a-420a-ba97-f161c5683757" (UID: "0f801e56-7e6a-420a-ba97-f161c5683757"). InnerVolumeSpecName "kube-api-access-j7kbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.929162 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-config-data" (OuterVolumeSpecName: "config-data") pod "0f801e56-7e6a-420a-ba97-f161c5683757" (UID: "0f801e56-7e6a-420a-ba97-f161c5683757"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:13 crc kubenswrapper[4733]: I1204 19:16:13.930588 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f801e56-7e6a-420a-ba97-f161c5683757" (UID: "0f801e56-7e6a-420a-ba97-f161c5683757"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.002192 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.002241 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-scripts\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.002291 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-config-data\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.002352 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6vmg\" (UniqueName: \"kubernetes.io/projected/b9d10c32-2f7a-4306-82cc-f84d893f2045-kube-api-access-r6vmg\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.002522 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.002538 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7kbc\" (UniqueName: \"kubernetes.io/projected/0f801e56-7e6a-420a-ba97-f161c5683757-kube-api-access-j7kbc\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.002552 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f801e56-7e6a-420a-ba97-f161c5683757-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.005980 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f801e56-7e6a-420a-ba97-f161c5683757" containerID="8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca" exitCode=0 Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.006025 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f801e56-7e6a-420a-ba97-f161c5683757","Type":"ContainerDied","Data":"8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca"} Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.006098 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0f801e56-7e6a-420a-ba97-f161c5683757","Type":"ContainerDied","Data":"5418f842e3d701017e9bd3f43800bb53839e6f4db339ada0e9a4b1b086a85dee"} Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.006095 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.006115 4733 scope.go:117] "RemoveContainer" containerID="8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.006640 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-scripts\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.007243 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.019578 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6vmg\" (UniqueName: \"kubernetes.io/projected/b9d10c32-2f7a-4306-82cc-f84d893f2045-kube-api-access-r6vmg\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.022209 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-config-data\") pod \"nova-cell1-cell-mapping-6v6gj\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.046572 4733 scope.go:117] "RemoveContainer" containerID="8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca" Dec 04 19:16:14 crc kubenswrapper[4733]: E1204 19:16:14.056810 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca\": container with ID starting with 8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca not found: ID does not exist" containerID="8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.056855 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca"} err="failed to get container status \"8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca\": rpc error: code = NotFound desc = could not find container \"8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca\": container with ID starting with 8c529b4e0b88922c01711cff0228142d7f3f844b75a42258ea14589bb07addca not found: ID does not exist" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.056922 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.079973 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.088072 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.089569 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.092561 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.101335 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.137935 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.207897 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqgv2\" (UniqueName: \"kubernetes.io/projected/0b22e9ab-c0c7-41a4-a42c-008df49460f9-kube-api-access-pqgv2\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.207961 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.207996 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-config-data\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.310145 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqgv2\" (UniqueName: \"kubernetes.io/projected/0b22e9ab-c0c7-41a4-a42c-008df49460f9-kube-api-access-pqgv2\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.310198 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.310233 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-config-data\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.316457 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-config-data\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.317689 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.331368 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqgv2\" (UniqueName: \"kubernetes.io/projected/0b22e9ab-c0c7-41a4-a42c-008df49460f9-kube-api-access-pqgv2\") pod \"nova-scheduler-0\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.357356 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f801e56-7e6a-420a-ba97-f161c5683757" path="/var/lib/kubelet/pods/0f801e56-7e6a-420a-ba97-f161c5683757/volumes" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.358490 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c881d71-9509-4330-a864-76b596c5b5a6" path="/var/lib/kubelet/pods/9c881d71-9509-4330-a864-76b596c5b5a6/volumes" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.416852 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.437870 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6v6gj"] Dec 04 19:16:14 crc kubenswrapper[4733]: I1204 19:16:14.874618 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:14 crc kubenswrapper[4733]: W1204 19:16:14.877409 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b22e9ab_c0c7_41a4_a42c_008df49460f9.slice/crio-1688be4b90bb03013f393baf568b693e25bb6ce2e1d3e451cd22f453cc40ce7c WatchSource:0}: Error finding container 1688be4b90bb03013f393baf568b693e25bb6ce2e1d3e451cd22f453cc40ce7c: Status 404 returned error can't find the container with id 1688be4b90bb03013f393baf568b693e25bb6ce2e1d3e451cd22f453cc40ce7c Dec 04 19:16:15 crc kubenswrapper[4733]: I1204 19:16:15.017650 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0b22e9ab-c0c7-41a4-a42c-008df49460f9","Type":"ContainerStarted","Data":"1688be4b90bb03013f393baf568b693e25bb6ce2e1d3e451cd22f453cc40ce7c"} Dec 04 19:16:15 crc kubenswrapper[4733]: I1204 19:16:15.020305 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6v6gj" event={"ID":"b9d10c32-2f7a-4306-82cc-f84d893f2045","Type":"ContainerStarted","Data":"bbab5feb0fd684ae57feacfea6fd3f334d1b1e5d5e8e136762c0eaff06c00982"} Dec 04 19:16:15 crc kubenswrapper[4733]: I1204 19:16:15.020340 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6v6gj" event={"ID":"b9d10c32-2f7a-4306-82cc-f84d893f2045","Type":"ContainerStarted","Data":"78a16129e62d9d3689757eb15227106657b2b59ac7f6f696a85895f19968dbae"} Dec 04 19:16:15 crc kubenswrapper[4733]: I1204 19:16:15.051044 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-6v6gj" podStartSLOduration=2.051023732 podStartE2EDuration="2.051023732s" podCreationTimestamp="2025-12-04 19:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:15.041459173 +0000 UTC m=+5836.996820219" watchObservedRunningTime="2025-12-04 19:16:15.051023732 +0000 UTC m=+5837.006384788" Dec 04 19:16:15 crc kubenswrapper[4733]: I1204 19:16:15.409074 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 19:16:15 crc kubenswrapper[4733]: I1204 19:16:15.409194 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 19:16:16 crc kubenswrapper[4733]: I1204 19:16:16.031608 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0b22e9ab-c0c7-41a4-a42c-008df49460f9","Type":"ContainerStarted","Data":"f6c64785aae97382dacf9abf52efac3d00c7363c439e7b2bef63f6a649f88429"} Dec 04 19:16:16 crc kubenswrapper[4733]: I1204 19:16:16.056728 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.056711702 podStartE2EDuration="2.056711702s" podCreationTimestamp="2025-12-04 19:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:16.046775133 +0000 UTC m=+5838.002136179" watchObservedRunningTime="2025-12-04 19:16:16.056711702 +0000 UTC m=+5838.012072758" Dec 04 19:16:19 crc kubenswrapper[4733]: I1204 19:16:19.418523 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 19:16:20 crc kubenswrapper[4733]: I1204 19:16:20.084978 4733 generic.go:334] "Generic (PLEG): container finished" podID="b9d10c32-2f7a-4306-82cc-f84d893f2045" containerID="bbab5feb0fd684ae57feacfea6fd3f334d1b1e5d5e8e136762c0eaff06c00982" exitCode=0 Dec 04 19:16:20 crc kubenswrapper[4733]: I1204 19:16:20.085359 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6v6gj" event={"ID":"b9d10c32-2f7a-4306-82cc-f84d893f2045","Type":"ContainerDied","Data":"bbab5feb0fd684ae57feacfea6fd3f334d1b1e5d5e8e136762c0eaff06c00982"} Dec 04 19:16:20 crc kubenswrapper[4733]: I1204 19:16:20.362630 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 19:16:20 crc kubenswrapper[4733]: I1204 19:16:20.362689 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 19:16:20 crc kubenswrapper[4733]: I1204 19:16:20.410754 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 19:16:20 crc kubenswrapper[4733]: I1204 19:16:20.410852 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.444984 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.445343 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.459590 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.528942 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.529300 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.558915 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-scripts\") pod \"b9d10c32-2f7a-4306-82cc-f84d893f2045\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.559017 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6vmg\" (UniqueName: \"kubernetes.io/projected/b9d10c32-2f7a-4306-82cc-f84d893f2045-kube-api-access-r6vmg\") pod \"b9d10c32-2f7a-4306-82cc-f84d893f2045\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.559109 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-combined-ca-bundle\") pod \"b9d10c32-2f7a-4306-82cc-f84d893f2045\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.559195 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-config-data\") pod \"b9d10c32-2f7a-4306-82cc-f84d893f2045\" (UID: \"b9d10c32-2f7a-4306-82cc-f84d893f2045\") " Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.564465 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9d10c32-2f7a-4306-82cc-f84d893f2045-kube-api-access-r6vmg" (OuterVolumeSpecName: "kube-api-access-r6vmg") pod "b9d10c32-2f7a-4306-82cc-f84d893f2045" (UID: "b9d10c32-2f7a-4306-82cc-f84d893f2045"). InnerVolumeSpecName "kube-api-access-r6vmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.564884 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-scripts" (OuterVolumeSpecName: "scripts") pod "b9d10c32-2f7a-4306-82cc-f84d893f2045" (UID: "b9d10c32-2f7a-4306-82cc-f84d893f2045"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.607425 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-config-data" (OuterVolumeSpecName: "config-data") pod "b9d10c32-2f7a-4306-82cc-f84d893f2045" (UID: "b9d10c32-2f7a-4306-82cc-f84d893f2045"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.607754 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9d10c32-2f7a-4306-82cc-f84d893f2045" (UID: "b9d10c32-2f7a-4306-82cc-f84d893f2045"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.661480 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.661530 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.661544 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9d10c32-2f7a-4306-82cc-f84d893f2045-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:21 crc kubenswrapper[4733]: I1204 19:16:21.661557 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6vmg\" (UniqueName: \"kubernetes.io/projected/b9d10c32-2f7a-4306-82cc-f84d893f2045-kube-api-access-r6vmg\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.098719 4733 scope.go:117] "RemoveContainer" containerID="b6631a3850f93db5a752c124cb517422fe9d4abd89bb55b8ddbaced614693bfc" Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.106734 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6v6gj" event={"ID":"b9d10c32-2f7a-4306-82cc-f84d893f2045","Type":"ContainerDied","Data":"78a16129e62d9d3689757eb15227106657b2b59ac7f6f696a85895f19968dbae"} Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.106990 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78a16129e62d9d3689757eb15227106657b2b59ac7f6f696a85895f19968dbae" Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.106781 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6v6gj" Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.300530 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.300776 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-log" containerID="cri-o://2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86" gracePeriod=30 Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.300870 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-api" containerID="cri-o://2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a" gracePeriod=30 Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.325166 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.325453 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0b22e9ab-c0c7-41a4-a42c-008df49460f9" containerName="nova-scheduler-scheduler" containerID="cri-o://f6c64785aae97382dacf9abf52efac3d00c7363c439e7b2bef63f6a649f88429" gracePeriod=30 Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.393722 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.394349 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-log" containerID="cri-o://b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0" gracePeriod=30 Dec 04 19:16:22 crc kubenswrapper[4733]: I1204 19:16:22.394610 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-metadata" containerID="cri-o://336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1" gracePeriod=30 Dec 04 19:16:23 crc kubenswrapper[4733]: I1204 19:16:23.154386 4733 generic.go:334] "Generic (PLEG): container finished" podID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerID="2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86" exitCode=143 Dec 04 19:16:23 crc kubenswrapper[4733]: I1204 19:16:23.154500 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d647bcd8-cc11-4c59-829b-4a4f1a568d1f","Type":"ContainerDied","Data":"2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86"} Dec 04 19:16:23 crc kubenswrapper[4733]: I1204 19:16:23.174052 4733 generic.go:334] "Generic (PLEG): container finished" podID="0637413e-e16b-40dc-915e-09985282c533" containerID="b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0" exitCode=143 Dec 04 19:16:23 crc kubenswrapper[4733]: I1204 19:16:23.174132 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0637413e-e16b-40dc-915e-09985282c533","Type":"ContainerDied","Data":"b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0"} Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.204515 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.205695 4733 generic.go:334] "Generic (PLEG): container finished" podID="0b22e9ab-c0c7-41a4-a42c-008df49460f9" containerID="f6c64785aae97382dacf9abf52efac3d00c7363c439e7b2bef63f6a649f88429" exitCode=0 Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.205749 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0b22e9ab-c0c7-41a4-a42c-008df49460f9","Type":"ContainerDied","Data":"f6c64785aae97382dacf9abf52efac3d00c7363c439e7b2bef63f6a649f88429"} Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.207280 4733 generic.go:334] "Generic (PLEG): container finished" podID="0637413e-e16b-40dc-915e-09985282c533" containerID="336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1" exitCode=0 Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.207305 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0637413e-e16b-40dc-915e-09985282c533","Type":"ContainerDied","Data":"336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1"} Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.207323 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0637413e-e16b-40dc-915e-09985282c533","Type":"ContainerDied","Data":"a2cb75a273ba1fe89f826ce6ff4c77f88ee475813bcad4bfaf1782982f0451a7"} Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.207342 4733 scope.go:117] "RemoveContainer" containerID="336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.207459 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.246404 4733 scope.go:117] "RemoveContainer" containerID="b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.251885 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-config-data\") pod \"0637413e-e16b-40dc-915e-09985282c533\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.251968 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0637413e-e16b-40dc-915e-09985282c533-logs\") pod \"0637413e-e16b-40dc-915e-09985282c533\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.252002 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-combined-ca-bundle\") pod \"0637413e-e16b-40dc-915e-09985282c533\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.252052 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwr4n\" (UniqueName: \"kubernetes.io/projected/0637413e-e16b-40dc-915e-09985282c533-kube-api-access-fwr4n\") pod \"0637413e-e16b-40dc-915e-09985282c533\" (UID: \"0637413e-e16b-40dc-915e-09985282c533\") " Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.253455 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0637413e-e16b-40dc-915e-09985282c533-logs" (OuterVolumeSpecName: "logs") pod "0637413e-e16b-40dc-915e-09985282c533" (UID: "0637413e-e16b-40dc-915e-09985282c533"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.259347 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0637413e-e16b-40dc-915e-09985282c533-kube-api-access-fwr4n" (OuterVolumeSpecName: "kube-api-access-fwr4n") pod "0637413e-e16b-40dc-915e-09985282c533" (UID: "0637413e-e16b-40dc-915e-09985282c533"). InnerVolumeSpecName "kube-api-access-fwr4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.279058 4733 scope.go:117] "RemoveContainer" containerID="336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1" Dec 04 19:16:26 crc kubenswrapper[4733]: E1204 19:16:26.279567 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1\": container with ID starting with 336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1 not found: ID does not exist" containerID="336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.279619 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1"} err="failed to get container status \"336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1\": rpc error: code = NotFound desc = could not find container \"336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1\": container with ID starting with 336784ce87ee69e4d6f49dcbc667d2f2011951fe0f0045115e47131df32d1fa1 not found: ID does not exist" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.279654 4733 scope.go:117] "RemoveContainer" containerID="b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0" Dec 04 19:16:26 crc kubenswrapper[4733]: E1204 19:16:26.280136 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0\": container with ID starting with b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0 not found: ID does not exist" containerID="b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.280170 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0"} err="failed to get container status \"b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0\": rpc error: code = NotFound desc = could not find container \"b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0\": container with ID starting with b86adfc319090774ecfe1ec2a99ae6bfac7b2faa7840b746dc509a171ae652e0 not found: ID does not exist" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.284463 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0637413e-e16b-40dc-915e-09985282c533" (UID: "0637413e-e16b-40dc-915e-09985282c533"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.286860 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-config-data" (OuterVolumeSpecName: "config-data") pod "0637413e-e16b-40dc-915e-09985282c533" (UID: "0637413e-e16b-40dc-915e-09985282c533"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.335709 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:16:26 crc kubenswrapper[4733]: E1204 19:16:26.336146 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.354551 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwr4n\" (UniqueName: \"kubernetes.io/projected/0637413e-e16b-40dc-915e-09985282c533-kube-api-access-fwr4n\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.354596 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.354611 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0637413e-e16b-40dc-915e-09985282c533-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.354625 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0637413e-e16b-40dc-915e-09985282c533-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.409646 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.455457 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqgv2\" (UniqueName: \"kubernetes.io/projected/0b22e9ab-c0c7-41a4-a42c-008df49460f9-kube-api-access-pqgv2\") pod \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.455529 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-config-data\") pod \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.455561 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-combined-ca-bundle\") pod \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\" (UID: \"0b22e9ab-c0c7-41a4-a42c-008df49460f9\") " Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.458754 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b22e9ab-c0c7-41a4-a42c-008df49460f9-kube-api-access-pqgv2" (OuterVolumeSpecName: "kube-api-access-pqgv2") pod "0b22e9ab-c0c7-41a4-a42c-008df49460f9" (UID: "0b22e9ab-c0c7-41a4-a42c-008df49460f9"). InnerVolumeSpecName "kube-api-access-pqgv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.478053 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-config-data" (OuterVolumeSpecName: "config-data") pod "0b22e9ab-c0c7-41a4-a42c-008df49460f9" (UID: "0b22e9ab-c0c7-41a4-a42c-008df49460f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.478822 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b22e9ab-c0c7-41a4-a42c-008df49460f9" (UID: "0b22e9ab-c0c7-41a4-a42c-008df49460f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.536656 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.549590 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.559582 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.559613 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b22e9ab-c0c7-41a4-a42c-008df49460f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.559623 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqgv2\" (UniqueName: \"kubernetes.io/projected/0b22e9ab-c0c7-41a4-a42c-008df49460f9-kube-api-access-pqgv2\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562113 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:26 crc kubenswrapper[4733]: E1204 19:16:26.562464 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-metadata" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562480 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-metadata" Dec 04 19:16:26 crc kubenswrapper[4733]: E1204 19:16:26.562490 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b22e9ab-c0c7-41a4-a42c-008df49460f9" containerName="nova-scheduler-scheduler" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562496 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b22e9ab-c0c7-41a4-a42c-008df49460f9" containerName="nova-scheduler-scheduler" Dec 04 19:16:26 crc kubenswrapper[4733]: E1204 19:16:26.562505 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9d10c32-2f7a-4306-82cc-f84d893f2045" containerName="nova-manage" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562510 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9d10c32-2f7a-4306-82cc-f84d893f2045" containerName="nova-manage" Dec 04 19:16:26 crc kubenswrapper[4733]: E1204 19:16:26.562543 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-log" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562549 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-log" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562696 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9d10c32-2f7a-4306-82cc-f84d893f2045" containerName="nova-manage" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562728 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b22e9ab-c0c7-41a4-a42c-008df49460f9" containerName="nova-scheduler-scheduler" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562750 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-log" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.562765 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0637413e-e16b-40dc-915e-09985282c533" containerName="nova-metadata-metadata" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.563606 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.566592 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.572073 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.660472 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b65d76-f6f6-4840-961a-e0878c140d78-logs\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.660611 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.660634 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-config-data\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.660670 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqh7l\" (UniqueName: \"kubernetes.io/projected/34b65d76-f6f6-4840-961a-e0878c140d78-kube-api-access-nqh7l\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.762635 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.762687 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-config-data\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.762746 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqh7l\" (UniqueName: \"kubernetes.io/projected/34b65d76-f6f6-4840-961a-e0878c140d78-kube-api-access-nqh7l\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.762848 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b65d76-f6f6-4840-961a-e0878c140d78-logs\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.764847 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b65d76-f6f6-4840-961a-e0878c140d78-logs\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.767330 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.767779 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-config-data\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.781925 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqh7l\" (UniqueName: \"kubernetes.io/projected/34b65d76-f6f6-4840-961a-e0878c140d78-kube-api-access-nqh7l\") pod \"nova-metadata-0\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " pod="openstack/nova-metadata-0" Dec 04 19:16:26 crc kubenswrapper[4733]: I1204 19:16:26.878142 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.019775 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.076488 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-config-data\") pod \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.077234 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-combined-ca-bundle\") pod \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.077320 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-logs\") pod \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.077384 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgtmn\" (UniqueName: \"kubernetes.io/projected/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-kube-api-access-rgtmn\") pod \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\" (UID: \"d647bcd8-cc11-4c59-829b-4a4f1a568d1f\") " Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.077770 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-logs" (OuterVolumeSpecName: "logs") pod "d647bcd8-cc11-4c59-829b-4a4f1a568d1f" (UID: "d647bcd8-cc11-4c59-829b-4a4f1a568d1f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.078041 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.083987 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-kube-api-access-rgtmn" (OuterVolumeSpecName: "kube-api-access-rgtmn") pod "d647bcd8-cc11-4c59-829b-4a4f1a568d1f" (UID: "d647bcd8-cc11-4c59-829b-4a4f1a568d1f"). InnerVolumeSpecName "kube-api-access-rgtmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.100341 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-config-data" (OuterVolumeSpecName: "config-data") pod "d647bcd8-cc11-4c59-829b-4a4f1a568d1f" (UID: "d647bcd8-cc11-4c59-829b-4a4f1a568d1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.111071 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d647bcd8-cc11-4c59-829b-4a4f1a568d1f" (UID: "d647bcd8-cc11-4c59-829b-4a4f1a568d1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.180246 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.180291 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.180308 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgtmn\" (UniqueName: \"kubernetes.io/projected/d647bcd8-cc11-4c59-829b-4a4f1a568d1f-kube-api-access-rgtmn\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.223956 4733 generic.go:334] "Generic (PLEG): container finished" podID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerID="2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a" exitCode=0 Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.224002 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d647bcd8-cc11-4c59-829b-4a4f1a568d1f","Type":"ContainerDied","Data":"2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a"} Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.224036 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.224054 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d647bcd8-cc11-4c59-829b-4a4f1a568d1f","Type":"ContainerDied","Data":"8eb4f0fd5f99dc569104ae1186dff6157d87dc706cf59b5382e330597c66cd32"} Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.224073 4733 scope.go:117] "RemoveContainer" containerID="2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.225446 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0b22e9ab-c0c7-41a4-a42c-008df49460f9","Type":"ContainerDied","Data":"1688be4b90bb03013f393baf568b693e25bb6ce2e1d3e451cd22f453cc40ce7c"} Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.225493 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.256371 4733 scope.go:117] "RemoveContainer" containerID="2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.295495 4733 scope.go:117] "RemoveContainer" containerID="2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a" Dec 04 19:16:27 crc kubenswrapper[4733]: E1204 19:16:27.296656 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a\": container with ID starting with 2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a not found: ID does not exist" containerID="2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.296708 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a"} err="failed to get container status \"2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a\": rpc error: code = NotFound desc = could not find container \"2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a\": container with ID starting with 2a464947a001ab740ddeaf4096ff6fd91968600c557e202d811d33e841c44e4a not found: ID does not exist" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.296755 4733 scope.go:117] "RemoveContainer" containerID="2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86" Dec 04 19:16:27 crc kubenswrapper[4733]: E1204 19:16:27.297160 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86\": container with ID starting with 2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86 not found: ID does not exist" containerID="2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.297201 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86"} err="failed to get container status \"2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86\": rpc error: code = NotFound desc = could not find container \"2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86\": container with ID starting with 2909599b0694ae86ba0e4abbdbf25f3d6f61a9bffc05fcb9bae03ad8ed10ae86 not found: ID does not exist" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.297220 4733 scope.go:117] "RemoveContainer" containerID="f6c64785aae97382dacf9abf52efac3d00c7363c439e7b2bef63f6a649f88429" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.307269 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.323647 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.335950 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.346713 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: W1204 19:16:27.349027 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34b65d76_f6f6_4840_961a_e0878c140d78.slice/crio-68e8a0517089fe7df20037e8ad4ab18fb28834ce255743080f1fb203d3695fcd WatchSource:0}: Error finding container 68e8a0517089fe7df20037e8ad4ab18fb28834ce255743080f1fb203d3695fcd: Status 404 returned error can't find the container with id 68e8a0517089fe7df20037e8ad4ab18fb28834ce255743080f1fb203d3695fcd Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.354623 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: E1204 19:16:27.355069 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-log" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.355092 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-log" Dec 04 19:16:27 crc kubenswrapper[4733]: E1204 19:16:27.355116 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-api" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.355123 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-api" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.355277 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-api" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.355292 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" containerName="nova-api-log" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.356340 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.358496 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.362541 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.371298 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.376233 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.378516 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.381021 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.385561 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-config-data\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.385622 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrmc6\" (UniqueName: \"kubernetes.io/projected/affe754f-50b9-4d77-9df2-027670584ddf-kube-api-access-xrmc6\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.385707 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affe754f-50b9-4d77-9df2-027670584ddf-logs\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.385916 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.389154 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.487301 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.487351 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.487405 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw8bl\" (UniqueName: \"kubernetes.io/projected/1099964c-7106-4254-9bd5-551e2c73f1b8-kube-api-access-bw8bl\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.487443 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-config-data\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.487465 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrmc6\" (UniqueName: \"kubernetes.io/projected/affe754f-50b9-4d77-9df2-027670584ddf-kube-api-access-xrmc6\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.487504 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affe754f-50b9-4d77-9df2-027670584ddf-logs\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.487548 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-config-data\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.488145 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affe754f-50b9-4d77-9df2-027670584ddf-logs\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.491173 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.491908 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-config-data\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.506937 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrmc6\" (UniqueName: \"kubernetes.io/projected/affe754f-50b9-4d77-9df2-027670584ddf-kube-api-access-xrmc6\") pod \"nova-api-0\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.588544 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-config-data\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.588619 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.588653 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw8bl\" (UniqueName: \"kubernetes.io/projected/1099964c-7106-4254-9bd5-551e2c73f1b8-kube-api-access-bw8bl\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.592702 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.593327 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-config-data\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.606972 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw8bl\" (UniqueName: \"kubernetes.io/projected/1099964c-7106-4254-9bd5-551e2c73f1b8-kube-api-access-bw8bl\") pod \"nova-scheduler-0\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " pod="openstack/nova-scheduler-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.675697 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:16:27 crc kubenswrapper[4733]: I1204 19:16:27.695301 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.202416 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:16:28 crc kubenswrapper[4733]: W1204 19:16:28.208905 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaffe754f_50b9_4d77_9df2_027670584ddf.slice/crio-0310c93c1ce1267fcb27f958c805506cea0240d111bca49804923b712abefd1b WatchSource:0}: Error finding container 0310c93c1ce1267fcb27f958c805506cea0240d111bca49804923b712abefd1b: Status 404 returned error can't find the container with id 0310c93c1ce1267fcb27f958c805506cea0240d111bca49804923b712abefd1b Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.243331 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"affe754f-50b9-4d77-9df2-027670584ddf","Type":"ContainerStarted","Data":"0310c93c1ce1267fcb27f958c805506cea0240d111bca49804923b712abefd1b"} Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.253443 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b65d76-f6f6-4840-961a-e0878c140d78","Type":"ContainerStarted","Data":"8cf1b4a2a70c8036c7b4eda3f909c4d2fadfe89bf04ba4ef470466c867cc3f95"} Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.253509 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b65d76-f6f6-4840-961a-e0878c140d78","Type":"ContainerStarted","Data":"18f4a09d21c8d0312c61833161ae0e52a65bfeb581a657ba4050c6dba468083b"} Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.253529 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b65d76-f6f6-4840-961a-e0878c140d78","Type":"ContainerStarted","Data":"68e8a0517089fe7df20037e8ad4ab18fb28834ce255743080f1fb203d3695fcd"} Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.266327 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.287409 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.287388903 podStartE2EDuration="2.287388903s" podCreationTimestamp="2025-12-04 19:16:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:28.277514696 +0000 UTC m=+5850.232875752" watchObservedRunningTime="2025-12-04 19:16:28.287388903 +0000 UTC m=+5850.242749959" Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.365463 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0637413e-e16b-40dc-915e-09985282c533" path="/var/lib/kubelet/pods/0637413e-e16b-40dc-915e-09985282c533/volumes" Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.366278 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b22e9ab-c0c7-41a4-a42c-008df49460f9" path="/var/lib/kubelet/pods/0b22e9ab-c0c7-41a4-a42c-008df49460f9/volumes" Dec 04 19:16:28 crc kubenswrapper[4733]: I1204 19:16:28.367107 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d647bcd8-cc11-4c59-829b-4a4f1a568d1f" path="/var/lib/kubelet/pods/d647bcd8-cc11-4c59-829b-4a4f1a568d1f/volumes" Dec 04 19:16:29 crc kubenswrapper[4733]: I1204 19:16:29.272452 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1099964c-7106-4254-9bd5-551e2c73f1b8","Type":"ContainerStarted","Data":"5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38"} Dec 04 19:16:29 crc kubenswrapper[4733]: I1204 19:16:29.272845 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1099964c-7106-4254-9bd5-551e2c73f1b8","Type":"ContainerStarted","Data":"303e67224f294ebbd3dad9a9622475bd6b7d88916d61011e5f7d80f8596a677d"} Dec 04 19:16:29 crc kubenswrapper[4733]: I1204 19:16:29.276251 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"affe754f-50b9-4d77-9df2-027670584ddf","Type":"ContainerStarted","Data":"c13f6bfc3ac4e3925c706789f94e51fb144bf32d337a9c4f2409c318e518cb43"} Dec 04 19:16:29 crc kubenswrapper[4733]: I1204 19:16:29.276301 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"affe754f-50b9-4d77-9df2-027670584ddf","Type":"ContainerStarted","Data":"6393ac159e85a10e047d0ef5367d40a9156fd4600cf811b07ff519153ce6d797"} Dec 04 19:16:29 crc kubenswrapper[4733]: I1204 19:16:29.296768 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.296747903 podStartE2EDuration="2.296747903s" podCreationTimestamp="2025-12-04 19:16:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:29.29256654 +0000 UTC m=+5851.247927586" watchObservedRunningTime="2025-12-04 19:16:29.296747903 +0000 UTC m=+5851.252108949" Dec 04 19:16:29 crc kubenswrapper[4733]: I1204 19:16:29.321696 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.3216813370000002 podStartE2EDuration="2.321681337s" podCreationTimestamp="2025-12-04 19:16:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:29.318605324 +0000 UTC m=+5851.273966370" watchObservedRunningTime="2025-12-04 19:16:29.321681337 +0000 UTC m=+5851.277042383" Dec 04 19:16:31 crc kubenswrapper[4733]: I1204 19:16:31.878608 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 19:16:31 crc kubenswrapper[4733]: I1204 19:16:31.879047 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 19:16:32 crc kubenswrapper[4733]: I1204 19:16:32.696102 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 19:16:36 crc kubenswrapper[4733]: I1204 19:16:36.878617 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 19:16:36 crc kubenswrapper[4733]: I1204 19:16:36.879741 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 19:16:37 crc kubenswrapper[4733]: I1204 19:16:37.335888 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:16:37 crc kubenswrapper[4733]: E1204 19:16:37.336199 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:16:37 crc kubenswrapper[4733]: I1204 19:16:37.676202 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 19:16:37 crc kubenswrapper[4733]: I1204 19:16:37.676595 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 19:16:37 crc kubenswrapper[4733]: I1204 19:16:37.696266 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 19:16:37 crc kubenswrapper[4733]: I1204 19:16:37.737276 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 19:16:37 crc kubenswrapper[4733]: I1204 19:16:37.961049 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.77:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:16:37 crc kubenswrapper[4733]: I1204 19:16:37.961059 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.77:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:16:38 crc kubenswrapper[4733]: I1204 19:16:38.406560 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 19:16:38 crc kubenswrapper[4733]: I1204 19:16:38.760002 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.78:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:16:38 crc kubenswrapper[4733]: I1204 19:16:38.760008 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.78:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:16:46 crc kubenswrapper[4733]: I1204 19:16:46.883033 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 19:16:46 crc kubenswrapper[4733]: I1204 19:16:46.886243 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 19:16:46 crc kubenswrapper[4733]: I1204 19:16:46.887283 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 19:16:47 crc kubenswrapper[4733]: I1204 19:16:47.462555 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 19:16:47 crc kubenswrapper[4733]: I1204 19:16:47.682399 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 19:16:47 crc kubenswrapper[4733]: I1204 19:16:47.683480 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 19:16:47 crc kubenswrapper[4733]: I1204 19:16:47.686382 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 19:16:47 crc kubenswrapper[4733]: I1204 19:16:47.687622 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.368157 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:16:48 crc kubenswrapper[4733]: E1204 19:16:48.368435 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.482240 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.486725 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.676719 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8675c4d89f-t6wg9"] Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.679506 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.702117 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8675c4d89f-t6wg9"] Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.832929 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-dns-svc\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.833049 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-nb\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.833087 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-config\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.833107 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48bfc\" (UniqueName: \"kubernetes.io/projected/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-kube-api-access-48bfc\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.833131 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-sb\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.934915 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-nb\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.935000 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-config\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.935036 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48bfc\" (UniqueName: \"kubernetes.io/projected/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-kube-api-access-48bfc\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.935079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-sb\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.935186 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-dns-svc\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.936284 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-dns-svc\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.937274 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-nb\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.938513 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-sb\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.938680 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-config\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:48 crc kubenswrapper[4733]: I1204 19:16:48.963762 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48bfc\" (UniqueName: \"kubernetes.io/projected/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-kube-api-access-48bfc\") pod \"dnsmasq-dns-8675c4d89f-t6wg9\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:49 crc kubenswrapper[4733]: I1204 19:16:49.023466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:49 crc kubenswrapper[4733]: I1204 19:16:49.517262 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8675c4d89f-t6wg9"] Dec 04 19:16:50 crc kubenswrapper[4733]: I1204 19:16:50.497625 4733 generic.go:334] "Generic (PLEG): container finished" podID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerID="c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2" exitCode=0 Dec 04 19:16:50 crc kubenswrapper[4733]: I1204 19:16:50.497783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" event={"ID":"cd9ddcea-451e-4d0b-94f4-821c3765eb0a","Type":"ContainerDied","Data":"c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2"} Dec 04 19:16:50 crc kubenswrapper[4733]: I1204 19:16:50.498987 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" event={"ID":"cd9ddcea-451e-4d0b-94f4-821c3765eb0a","Type":"ContainerStarted","Data":"f509e673ed3f416ab37c774ff33a3ac1782949925f9c477285ca84b6ec39fdcb"} Dec 04 19:16:51 crc kubenswrapper[4733]: I1204 19:16:51.513006 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" event={"ID":"cd9ddcea-451e-4d0b-94f4-821c3765eb0a","Type":"ContainerStarted","Data":"471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d"} Dec 04 19:16:51 crc kubenswrapper[4733]: I1204 19:16:51.513295 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:51 crc kubenswrapper[4733]: I1204 19:16:51.536366 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" podStartSLOduration=3.536349922 podStartE2EDuration="3.536349922s" podCreationTimestamp="2025-12-04 19:16:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:16:51.527185194 +0000 UTC m=+5873.482546240" watchObservedRunningTime="2025-12-04 19:16:51.536349922 +0000 UTC m=+5873.491710968" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.026008 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.134940 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f8b566c47-9l9wx"] Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.135532 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" podUID="ef390936-b761-4615-a1a9-517066f07745" containerName="dnsmasq-dns" containerID="cri-o://e2f457072bddbd8e832eb075ca553f01885aa55a715bbbaf7d273270a587d110" gracePeriod=10 Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.335426 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:16:59 crc kubenswrapper[4733]: E1204 19:16:59.335912 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.611069 4733 generic.go:334] "Generic (PLEG): container finished" podID="ef390936-b761-4615-a1a9-517066f07745" containerID="e2f457072bddbd8e832eb075ca553f01885aa55a715bbbaf7d273270a587d110" exitCode=0 Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.611212 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" event={"ID":"ef390936-b761-4615-a1a9-517066f07745","Type":"ContainerDied","Data":"e2f457072bddbd8e832eb075ca553f01885aa55a715bbbaf7d273270a587d110"} Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.611327 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" event={"ID":"ef390936-b761-4615-a1a9-517066f07745","Type":"ContainerDied","Data":"2af1ad118a45d6fdeffca88dd2699f8cb2d2f6b0e0379e7bb8e93d59696dcbd4"} Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.611338 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2af1ad118a45d6fdeffca88dd2699f8cb2d2f6b0e0379e7bb8e93d59696dcbd4" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.646574 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.845096 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcqbz\" (UniqueName: \"kubernetes.io/projected/ef390936-b761-4615-a1a9-517066f07745-kube-api-access-mcqbz\") pod \"ef390936-b761-4615-a1a9-517066f07745\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.845167 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-config\") pod \"ef390936-b761-4615-a1a9-517066f07745\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.845205 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-dns-svc\") pod \"ef390936-b761-4615-a1a9-517066f07745\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.845235 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-sb\") pod \"ef390936-b761-4615-a1a9-517066f07745\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.845318 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-nb\") pod \"ef390936-b761-4615-a1a9-517066f07745\" (UID: \"ef390936-b761-4615-a1a9-517066f07745\") " Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.861623 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef390936-b761-4615-a1a9-517066f07745-kube-api-access-mcqbz" (OuterVolumeSpecName: "kube-api-access-mcqbz") pod "ef390936-b761-4615-a1a9-517066f07745" (UID: "ef390936-b761-4615-a1a9-517066f07745"). InnerVolumeSpecName "kube-api-access-mcqbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.890484 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef390936-b761-4615-a1a9-517066f07745" (UID: "ef390936-b761-4615-a1a9-517066f07745"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.896673 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef390936-b761-4615-a1a9-517066f07745" (UID: "ef390936-b761-4615-a1a9-517066f07745"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.897643 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef390936-b761-4615-a1a9-517066f07745" (UID: "ef390936-b761-4615-a1a9-517066f07745"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.917588 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-config" (OuterVolumeSpecName: "config") pod "ef390936-b761-4615-a1a9-517066f07745" (UID: "ef390936-b761-4615-a1a9-517066f07745"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.947442 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcqbz\" (UniqueName: \"kubernetes.io/projected/ef390936-b761-4615-a1a9-517066f07745-kube-api-access-mcqbz\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.947478 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.947490 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.947502 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:16:59 crc kubenswrapper[4733]: I1204 19:16:59.947513 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef390936-b761-4615-a1a9-517066f07745-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.410066 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-lww2m"] Dec 04 19:17:00 crc kubenswrapper[4733]: E1204 19:17:00.410432 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef390936-b761-4615-a1a9-517066f07745" containerName="dnsmasq-dns" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.410445 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef390936-b761-4615-a1a9-517066f07745" containerName="dnsmasq-dns" Dec 04 19:17:00 crc kubenswrapper[4733]: E1204 19:17:00.410468 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef390936-b761-4615-a1a9-517066f07745" containerName="init" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.410474 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef390936-b761-4615-a1a9-517066f07745" containerName="init" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.410640 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef390936-b761-4615-a1a9-517066f07745" containerName="dnsmasq-dns" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.411439 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.418989 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lww2m"] Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.543099 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-7ec3-account-create-update-s4czg"] Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.545565 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.551489 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.552335 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7ec3-account-create-update-s4czg"] Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.564593 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-operator-scripts\") pod \"cinder-db-create-lww2m\" (UID: \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\") " pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.564641 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-operator-scripts\") pod \"cinder-7ec3-account-create-update-s4czg\" (UID: \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\") " pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.564719 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtz6h\" (UniqueName: \"kubernetes.io/projected/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-kube-api-access-wtz6h\") pod \"cinder-7ec3-account-create-update-s4czg\" (UID: \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\") " pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.564747 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rfjw\" (UniqueName: \"kubernetes.io/projected/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-kube-api-access-4rfjw\") pod \"cinder-db-create-lww2m\" (UID: \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\") " pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.619492 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f8b566c47-9l9wx" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.650877 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f8b566c47-9l9wx"] Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.662076 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f8b566c47-9l9wx"] Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.666277 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-operator-scripts\") pod \"cinder-db-create-lww2m\" (UID: \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\") " pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.666333 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-operator-scripts\") pod \"cinder-7ec3-account-create-update-s4czg\" (UID: \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\") " pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.666396 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtz6h\" (UniqueName: \"kubernetes.io/projected/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-kube-api-access-wtz6h\") pod \"cinder-7ec3-account-create-update-s4czg\" (UID: \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\") " pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.666422 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rfjw\" (UniqueName: \"kubernetes.io/projected/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-kube-api-access-4rfjw\") pod \"cinder-db-create-lww2m\" (UID: \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\") " pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.667092 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-operator-scripts\") pod \"cinder-db-create-lww2m\" (UID: \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\") " pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.667257 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-operator-scripts\") pod \"cinder-7ec3-account-create-update-s4czg\" (UID: \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\") " pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.684077 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rfjw\" (UniqueName: \"kubernetes.io/projected/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-kube-api-access-4rfjw\") pod \"cinder-db-create-lww2m\" (UID: \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\") " pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.685709 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtz6h\" (UniqueName: \"kubernetes.io/projected/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-kube-api-access-wtz6h\") pod \"cinder-7ec3-account-create-update-s4czg\" (UID: \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\") " pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.775690 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:00 crc kubenswrapper[4733]: I1204 19:17:00.864699 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:01 crc kubenswrapper[4733]: W1204 19:17:01.307992 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c0e84bb_07b0_4e93_b1eb_7170a0f35ec6.slice/crio-7eac028abb6a7f6408f23aac09b8aa7425caf9be5cf36a185c89e064fb95b09d WatchSource:0}: Error finding container 7eac028abb6a7f6408f23aac09b8aa7425caf9be5cf36a185c89e064fb95b09d: Status 404 returned error can't find the container with id 7eac028abb6a7f6408f23aac09b8aa7425caf9be5cf36a185c89e064fb95b09d Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.308248 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lww2m"] Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.410707 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7ec3-account-create-update-s4czg"] Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.608791 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dqtg2"] Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.611573 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.630508 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7ec3-account-create-update-s4czg" event={"ID":"a8ab8394-2a72-4898-a02b-a5bee23bdc6a","Type":"ContainerStarted","Data":"6bff829d438170545a05f0da01e56a6c678894249ab6223cab87c2ffbc649d5c"} Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.631427 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dqtg2"] Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.641368 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lww2m" event={"ID":"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6","Type":"ContainerStarted","Data":"7eac028abb6a7f6408f23aac09b8aa7425caf9be5cf36a185c89e064fb95b09d"} Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.785409 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-catalog-content\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.785484 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-utilities\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.785621 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qqcg\" (UniqueName: \"kubernetes.io/projected/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-kube-api-access-9qqcg\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.887821 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-utilities\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.887886 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qqcg\" (UniqueName: \"kubernetes.io/projected/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-kube-api-access-9qqcg\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.887977 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-catalog-content\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.888378 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-utilities\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.888575 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-catalog-content\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.907066 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qqcg\" (UniqueName: \"kubernetes.io/projected/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-kube-api-access-9qqcg\") pod \"community-operators-dqtg2\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:01 crc kubenswrapper[4733]: I1204 19:17:01.929985 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:02 crc kubenswrapper[4733]: I1204 19:17:02.351655 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef390936-b761-4615-a1a9-517066f07745" path="/var/lib/kubelet/pods/ef390936-b761-4615-a1a9-517066f07745/volumes" Dec 04 19:17:02 crc kubenswrapper[4733]: W1204 19:17:02.417351 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeed3d00a_7daa_40a4_bdfa_5b1ec6fe32e1.slice/crio-612eee00aa6d2945352994721e5282ed51a0276f9ff851f6f36dd968df66abf0 WatchSource:0}: Error finding container 612eee00aa6d2945352994721e5282ed51a0276f9ff851f6f36dd968df66abf0: Status 404 returned error can't find the container with id 612eee00aa6d2945352994721e5282ed51a0276f9ff851f6f36dd968df66abf0 Dec 04 19:17:02 crc kubenswrapper[4733]: I1204 19:17:02.420216 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dqtg2"] Dec 04 19:17:02 crc kubenswrapper[4733]: I1204 19:17:02.650158 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqtg2" event={"ID":"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1","Type":"ContainerStarted","Data":"17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297"} Dec 04 19:17:02 crc kubenswrapper[4733]: I1204 19:17:02.650204 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqtg2" event={"ID":"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1","Type":"ContainerStarted","Data":"612eee00aa6d2945352994721e5282ed51a0276f9ff851f6f36dd968df66abf0"} Dec 04 19:17:02 crc kubenswrapper[4733]: I1204 19:17:02.653810 4733 generic.go:334] "Generic (PLEG): container finished" podID="a8ab8394-2a72-4898-a02b-a5bee23bdc6a" containerID="d6c619b4a7aa7643365dc04fd2a29d4f7daaa747f2c733d1329e4f509ddffe97" exitCode=0 Dec 04 19:17:02 crc kubenswrapper[4733]: I1204 19:17:02.653934 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7ec3-account-create-update-s4czg" event={"ID":"a8ab8394-2a72-4898-a02b-a5bee23bdc6a","Type":"ContainerDied","Data":"d6c619b4a7aa7643365dc04fd2a29d4f7daaa747f2c733d1329e4f509ddffe97"} Dec 04 19:17:02 crc kubenswrapper[4733]: I1204 19:17:02.655877 4733 generic.go:334] "Generic (PLEG): container finished" podID="9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6" containerID="315c37c18e8368fb4803d151f7afbdb60d6429ff959bc96b58d3832fc05276d7" exitCode=0 Dec 04 19:17:02 crc kubenswrapper[4733]: I1204 19:17:02.655909 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lww2m" event={"ID":"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6","Type":"ContainerDied","Data":"315c37c18e8368fb4803d151f7afbdb60d6429ff959bc96b58d3832fc05276d7"} Dec 04 19:17:03 crc kubenswrapper[4733]: I1204 19:17:03.671575 4733 generic.go:334] "Generic (PLEG): container finished" podID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerID="17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297" exitCode=0 Dec 04 19:17:03 crc kubenswrapper[4733]: I1204 19:17:03.671792 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqtg2" event={"ID":"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1","Type":"ContainerDied","Data":"17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297"} Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.125452 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.135569 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtz6h\" (UniqueName: \"kubernetes.io/projected/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-kube-api-access-wtz6h\") pod \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\" (UID: \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\") " Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.135835 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-operator-scripts\") pod \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\" (UID: \"a8ab8394-2a72-4898-a02b-a5bee23bdc6a\") " Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.136950 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8ab8394-2a72-4898-a02b-a5bee23bdc6a" (UID: "a8ab8394-2a72-4898-a02b-a5bee23bdc6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.144311 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-kube-api-access-wtz6h" (OuterVolumeSpecName: "kube-api-access-wtz6h") pod "a8ab8394-2a72-4898-a02b-a5bee23bdc6a" (UID: "a8ab8394-2a72-4898-a02b-a5bee23bdc6a"). InnerVolumeSpecName "kube-api-access-wtz6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.230027 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.237153 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rfjw\" (UniqueName: \"kubernetes.io/projected/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-kube-api-access-4rfjw\") pod \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\" (UID: \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\") " Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.237276 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-operator-scripts\") pod \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\" (UID: \"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6\") " Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.237553 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.237565 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtz6h\" (UniqueName: \"kubernetes.io/projected/a8ab8394-2a72-4898-a02b-a5bee23bdc6a-kube-api-access-wtz6h\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.237966 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6" (UID: "9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.241113 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-kube-api-access-4rfjw" (OuterVolumeSpecName: "kube-api-access-4rfjw") pod "9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6" (UID: "9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6"). InnerVolumeSpecName "kube-api-access-4rfjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.338869 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rfjw\" (UniqueName: \"kubernetes.io/projected/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-kube-api-access-4rfjw\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.338916 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.687073 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqtg2" event={"ID":"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1","Type":"ContainerStarted","Data":"959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187"} Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.691621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7ec3-account-create-update-s4czg" event={"ID":"a8ab8394-2a72-4898-a02b-a5bee23bdc6a","Type":"ContainerDied","Data":"6bff829d438170545a05f0da01e56a6c678894249ab6223cab87c2ffbc649d5c"} Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.691702 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bff829d438170545a05f0da01e56a6c678894249ab6223cab87c2ffbc649d5c" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.691640 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7ec3-account-create-update-s4czg" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.695622 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lww2m" event={"ID":"9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6","Type":"ContainerDied","Data":"7eac028abb6a7f6408f23aac09b8aa7425caf9be5cf36a185c89e064fb95b09d"} Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.696070 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eac028abb6a7f6408f23aac09b8aa7425caf9be5cf36a185c89e064fb95b09d" Dec 04 19:17:04 crc kubenswrapper[4733]: I1204 19:17:04.695705 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lww2m" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.708973 4733 generic.go:334] "Generic (PLEG): container finished" podID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerID="959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187" exitCode=0 Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.709069 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqtg2" event={"ID":"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1","Type":"ContainerDied","Data":"959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187"} Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.861884 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-szccb"] Dec 04 19:17:05 crc kubenswrapper[4733]: E1204 19:17:05.862376 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6" containerName="mariadb-database-create" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.862399 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6" containerName="mariadb-database-create" Dec 04 19:17:05 crc kubenswrapper[4733]: E1204 19:17:05.862423 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ab8394-2a72-4898-a02b-a5bee23bdc6a" containerName="mariadb-account-create-update" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.862432 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ab8394-2a72-4898-a02b-a5bee23bdc6a" containerName="mariadb-account-create-update" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.862683 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8ab8394-2a72-4898-a02b-a5bee23bdc6a" containerName="mariadb-account-create-update" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.862708 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6" containerName="mariadb-database-create" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.863513 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.865031 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jlj47" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.866338 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.866376 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.873056 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-szccb"] Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.892276 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-db-sync-config-data\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.892686 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-config-data\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.892926 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/21a9d868-6c00-4586-b785-9efb6282c204-etc-machine-id\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.893116 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-combined-ca-bundle\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.893361 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcxcv\" (UniqueName: \"kubernetes.io/projected/21a9d868-6c00-4586-b785-9efb6282c204-kube-api-access-hcxcv\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.893454 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-scripts\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.995414 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-db-sync-config-data\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.995750 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-config-data\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.995853 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/21a9d868-6c00-4586-b785-9efb6282c204-etc-machine-id\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.995958 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-combined-ca-bundle\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.996018 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcxcv\" (UniqueName: \"kubernetes.io/projected/21a9d868-6c00-4586-b785-9efb6282c204-kube-api-access-hcxcv\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.996042 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-scripts\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:05 crc kubenswrapper[4733]: I1204 19:17:05.996915 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/21a9d868-6c00-4586-b785-9efb6282c204-etc-machine-id\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.002596 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-db-sync-config-data\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.008650 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-combined-ca-bundle\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.013866 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-config-data\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.014648 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-scripts\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.020667 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcxcv\" (UniqueName: \"kubernetes.io/projected/21a9d868-6c00-4586-b785-9efb6282c204-kube-api-access-hcxcv\") pod \"cinder-db-sync-szccb\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.219287 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:06 crc kubenswrapper[4733]: W1204 19:17:06.702397 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21a9d868_6c00_4586_b785_9efb6282c204.slice/crio-59ce27769178ebc3bd844a29d28dbdcf6a670d3ffbe672f7e3c5cf071c8bb064 WatchSource:0}: Error finding container 59ce27769178ebc3bd844a29d28dbdcf6a670d3ffbe672f7e3c5cf071c8bb064: Status 404 returned error can't find the container with id 59ce27769178ebc3bd844a29d28dbdcf6a670d3ffbe672f7e3c5cf071c8bb064 Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.703772 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-szccb"] Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.722352 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqtg2" event={"ID":"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1","Type":"ContainerStarted","Data":"117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2"} Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.726159 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-szccb" event={"ID":"21a9d868-6c00-4586-b785-9efb6282c204","Type":"ContainerStarted","Data":"59ce27769178ebc3bd844a29d28dbdcf6a670d3ffbe672f7e3c5cf071c8bb064"} Dec 04 19:17:06 crc kubenswrapper[4733]: I1204 19:17:06.752731 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dqtg2" podStartSLOduration=3.326769622 podStartE2EDuration="5.752711016s" podCreationTimestamp="2025-12-04 19:17:01 +0000 UTC" firstStartedPulling="2025-12-04 19:17:03.674151841 +0000 UTC m=+5885.629512907" lastFinishedPulling="2025-12-04 19:17:06.100093255 +0000 UTC m=+5888.055454301" observedRunningTime="2025-12-04 19:17:06.740348301 +0000 UTC m=+5888.695709367" watchObservedRunningTime="2025-12-04 19:17:06.752711016 +0000 UTC m=+5888.708072062" Dec 04 19:17:07 crc kubenswrapper[4733]: I1204 19:17:07.742137 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-szccb" event={"ID":"21a9d868-6c00-4586-b785-9efb6282c204","Type":"ContainerStarted","Data":"c7d4bf016c7eddc5eadae3a7b69d82654ea5c6e037dc4d50cc6ee226b99801bc"} Dec 04 19:17:07 crc kubenswrapper[4733]: I1204 19:17:07.769249 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-szccb" podStartSLOduration=2.769233 podStartE2EDuration="2.769233s" podCreationTimestamp="2025-12-04 19:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:07.765857828 +0000 UTC m=+5889.721218944" watchObservedRunningTime="2025-12-04 19:17:07.769233 +0000 UTC m=+5889.724594046" Dec 04 19:17:10 crc kubenswrapper[4733]: I1204 19:17:10.778596 4733 generic.go:334] "Generic (PLEG): container finished" podID="21a9d868-6c00-4586-b785-9efb6282c204" containerID="c7d4bf016c7eddc5eadae3a7b69d82654ea5c6e037dc4d50cc6ee226b99801bc" exitCode=0 Dec 04 19:17:10 crc kubenswrapper[4733]: I1204 19:17:10.778740 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-szccb" event={"ID":"21a9d868-6c00-4586-b785-9efb6282c204","Type":"ContainerDied","Data":"c7d4bf016c7eddc5eadae3a7b69d82654ea5c6e037dc4d50cc6ee226b99801bc"} Dec 04 19:17:11 crc kubenswrapper[4733]: I1204 19:17:11.930742 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:11 crc kubenswrapper[4733]: I1204 19:17:11.932996 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.020855 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.229355 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.319376 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcxcv\" (UniqueName: \"kubernetes.io/projected/21a9d868-6c00-4586-b785-9efb6282c204-kube-api-access-hcxcv\") pod \"21a9d868-6c00-4586-b785-9efb6282c204\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.319437 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-config-data\") pod \"21a9d868-6c00-4586-b785-9efb6282c204\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.319470 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-db-sync-config-data\") pod \"21a9d868-6c00-4586-b785-9efb6282c204\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.319485 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-scripts\") pod \"21a9d868-6c00-4586-b785-9efb6282c204\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.319577 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-combined-ca-bundle\") pod \"21a9d868-6c00-4586-b785-9efb6282c204\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.319601 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/21a9d868-6c00-4586-b785-9efb6282c204-etc-machine-id\") pod \"21a9d868-6c00-4586-b785-9efb6282c204\" (UID: \"21a9d868-6c00-4586-b785-9efb6282c204\") " Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.319759 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21a9d868-6c00-4586-b785-9efb6282c204-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "21a9d868-6c00-4586-b785-9efb6282c204" (UID: "21a9d868-6c00-4586-b785-9efb6282c204"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.320068 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/21a9d868-6c00-4586-b785-9efb6282c204-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.326046 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-scripts" (OuterVolumeSpecName: "scripts") pod "21a9d868-6c00-4586-b785-9efb6282c204" (UID: "21a9d868-6c00-4586-b785-9efb6282c204"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.326169 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "21a9d868-6c00-4586-b785-9efb6282c204" (UID: "21a9d868-6c00-4586-b785-9efb6282c204"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.326293 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21a9d868-6c00-4586-b785-9efb6282c204-kube-api-access-hcxcv" (OuterVolumeSpecName: "kube-api-access-hcxcv") pod "21a9d868-6c00-4586-b785-9efb6282c204" (UID: "21a9d868-6c00-4586-b785-9efb6282c204"). InnerVolumeSpecName "kube-api-access-hcxcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.336500 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:17:12 crc kubenswrapper[4733]: E1204 19:17:12.336690 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.354116 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21a9d868-6c00-4586-b785-9efb6282c204" (UID: "21a9d868-6c00-4586-b785-9efb6282c204"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.375691 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-config-data" (OuterVolumeSpecName: "config-data") pod "21a9d868-6c00-4586-b785-9efb6282c204" (UID: "21a9d868-6c00-4586-b785-9efb6282c204"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.422653 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.422737 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcxcv\" (UniqueName: \"kubernetes.io/projected/21a9d868-6c00-4586-b785-9efb6282c204-kube-api-access-hcxcv\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.422767 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.422790 4733 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.422853 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a9d868-6c00-4586-b785-9efb6282c204-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.840977 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-szccb" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.841045 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-szccb" event={"ID":"21a9d868-6c00-4586-b785-9efb6282c204","Type":"ContainerDied","Data":"59ce27769178ebc3bd844a29d28dbdcf6a670d3ffbe672f7e3c5cf071c8bb064"} Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.841079 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59ce27769178ebc3bd844a29d28dbdcf6a670d3ffbe672f7e3c5cf071c8bb064" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.923791 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:12 crc kubenswrapper[4733]: I1204 19:17:12.991986 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dqtg2"] Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.144237 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6"] Dec 04 19:17:13 crc kubenswrapper[4733]: E1204 19:17:13.144969 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a9d868-6c00-4586-b785-9efb6282c204" containerName="cinder-db-sync" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.144986 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a9d868-6c00-4586-b785-9efb6282c204" containerName="cinder-db-sync" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.145166 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a9d868-6c00-4586-b785-9efb6282c204" containerName="cinder-db-sync" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.146041 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.159834 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6"] Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.243228 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-dns-svc\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.243374 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnp9s\" (UniqueName: \"kubernetes.io/projected/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-kube-api-access-dnp9s\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.243449 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.243468 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-config\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.243615 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.345640 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnp9s\" (UniqueName: \"kubernetes.io/projected/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-kube-api-access-dnp9s\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.345726 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.345746 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-config\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.345765 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.345834 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-dns-svc\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.346623 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-dns-svc\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.347100 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.347228 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-config\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.347740 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.366538 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnp9s\" (UniqueName: \"kubernetes.io/projected/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-kube-api-access-dnp9s\") pod \"dnsmasq-dns-6d6cbfd8f5-9vdb6\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.478322 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.529727 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.531175 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.535096 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.535756 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.539104 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.539271 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jlj47" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.612525 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.653613 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.653676 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/464f9982-87db-4cc0-991d-83eb0148e61b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.653719 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk5h4\" (UniqueName: \"kubernetes.io/projected/464f9982-87db-4cc0-991d-83eb0148e61b-kube-api-access-vk5h4\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.653760 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/464f9982-87db-4cc0-991d-83eb0148e61b-logs\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.653777 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-scripts\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.653873 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.653889 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data-custom\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.755603 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/464f9982-87db-4cc0-991d-83eb0148e61b-logs\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.756023 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-scripts\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.756068 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/464f9982-87db-4cc0-991d-83eb0148e61b-logs\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.756134 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.756162 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data-custom\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.756215 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.756254 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/464f9982-87db-4cc0-991d-83eb0148e61b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.756301 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk5h4\" (UniqueName: \"kubernetes.io/projected/464f9982-87db-4cc0-991d-83eb0148e61b-kube-api-access-vk5h4\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.756446 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/464f9982-87db-4cc0-991d-83eb0148e61b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.760302 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data-custom\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.760617 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-scripts\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.763514 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.765683 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.774881 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk5h4\" (UniqueName: \"kubernetes.io/projected/464f9982-87db-4cc0-991d-83eb0148e61b-kube-api-access-vk5h4\") pod \"cinder-api-0\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " pod="openstack/cinder-api-0" Dec 04 19:17:13 crc kubenswrapper[4733]: I1204 19:17:13.942252 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 19:17:14 crc kubenswrapper[4733]: I1204 19:17:14.062873 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6"] Dec 04 19:17:14 crc kubenswrapper[4733]: I1204 19:17:14.417033 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:14 crc kubenswrapper[4733]: I1204 19:17:14.860075 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" containerID="a560b36ba69efa48249f7ec05a8d0ec39bac0e40408c719a17a56e177d29dc38" exitCode=0 Dec 04 19:17:14 crc kubenswrapper[4733]: I1204 19:17:14.860126 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" event={"ID":"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445","Type":"ContainerDied","Data":"a560b36ba69efa48249f7ec05a8d0ec39bac0e40408c719a17a56e177d29dc38"} Dec 04 19:17:14 crc kubenswrapper[4733]: I1204 19:17:14.860176 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" event={"ID":"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445","Type":"ContainerStarted","Data":"81d9c3b238954b6de13b3558f3341fcfe1838b9ccfdc8ff6865d4552e56d108d"} Dec 04 19:17:14 crc kubenswrapper[4733]: I1204 19:17:14.864724 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"464f9982-87db-4cc0-991d-83eb0148e61b","Type":"ContainerStarted","Data":"ddfc28af4ca6389988ec13a0270d776ef798255d7a17ae466c982b723241808f"} Dec 04 19:17:14 crc kubenswrapper[4733]: I1204 19:17:14.864867 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dqtg2" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerName="registry-server" containerID="cri-o://117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2" gracePeriod=2 Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.340171 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.507729 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-utilities\") pod \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.508309 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-catalog-content\") pod \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.508418 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qqcg\" (UniqueName: \"kubernetes.io/projected/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-kube-api-access-9qqcg\") pod \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\" (UID: \"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1\") " Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.508727 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-utilities" (OuterVolumeSpecName: "utilities") pod "eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" (UID: "eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.509207 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.516029 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-kube-api-access-9qqcg" (OuterVolumeSpecName: "kube-api-access-9qqcg") pod "eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" (UID: "eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1"). InnerVolumeSpecName "kube-api-access-9qqcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.576097 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" (UID: "eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.611067 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qqcg\" (UniqueName: \"kubernetes.io/projected/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-kube-api-access-9qqcg\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.611101 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.874455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"464f9982-87db-4cc0-991d-83eb0148e61b","Type":"ContainerStarted","Data":"ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d"} Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.876869 4733 generic.go:334] "Generic (PLEG): container finished" podID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerID="117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2" exitCode=0 Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.876940 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqtg2" event={"ID":"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1","Type":"ContainerDied","Data":"117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2"} Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.876970 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqtg2" event={"ID":"eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1","Type":"ContainerDied","Data":"612eee00aa6d2945352994721e5282ed51a0276f9ff851f6f36dd968df66abf0"} Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.876991 4733 scope.go:117] "RemoveContainer" containerID="117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.877132 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqtg2" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.880682 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" event={"ID":"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445","Type":"ContainerStarted","Data":"226de081d3f6abc10ef577d59c4e903368f4e18a3940aed03dceaea641ecb565"} Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.880852 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.900036 4733 scope.go:117] "RemoveContainer" containerID="959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.917934 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" podStartSLOduration=2.917911405 podStartE2EDuration="2.917911405s" podCreationTimestamp="2025-12-04 19:17:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:15.899361334 +0000 UTC m=+5897.854722390" watchObservedRunningTime="2025-12-04 19:17:15.917911405 +0000 UTC m=+5897.873272451" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.932636 4733 scope.go:117] "RemoveContainer" containerID="17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.934509 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dqtg2"] Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.945457 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dqtg2"] Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.982154 4733 scope.go:117] "RemoveContainer" containerID="117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2" Dec 04 19:17:15 crc kubenswrapper[4733]: E1204 19:17:15.983468 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2\": container with ID starting with 117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2 not found: ID does not exist" containerID="117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.983519 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2"} err="failed to get container status \"117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2\": rpc error: code = NotFound desc = could not find container \"117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2\": container with ID starting with 117a393fe356586cff90bc02e27b8a87df19c0a3bf8f8217b8ec0a6530013ae2 not found: ID does not exist" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.983564 4733 scope.go:117] "RemoveContainer" containerID="959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187" Dec 04 19:17:15 crc kubenswrapper[4733]: E1204 19:17:15.984052 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187\": container with ID starting with 959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187 not found: ID does not exist" containerID="959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.984104 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187"} err="failed to get container status \"959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187\": rpc error: code = NotFound desc = could not find container \"959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187\": container with ID starting with 959fa9689150e1fafcef45881fcffdd3ee081a48e1dd23a49897a3d4de7ef187 not found: ID does not exist" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.984136 4733 scope.go:117] "RemoveContainer" containerID="17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297" Dec 04 19:17:15 crc kubenswrapper[4733]: E1204 19:17:15.984680 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297\": container with ID starting with 17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297 not found: ID does not exist" containerID="17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297" Dec 04 19:17:15 crc kubenswrapper[4733]: I1204 19:17:15.984714 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297"} err="failed to get container status \"17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297\": rpc error: code = NotFound desc = could not find container \"17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297\": container with ID starting with 17c90871cabb52ad6390cbc6f80ce318e93917c3e093ce5efc1da52e502c6297 not found: ID does not exist" Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.131603 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.132196 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-log" containerID="cri-o://6393ac159e85a10e047d0ef5367d40a9156fd4600cf811b07ff519153ce6d797" gracePeriod=30 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.132303 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-api" containerID="cri-o://c13f6bfc3ac4e3925c706789f94e51fb144bf32d337a9c4f2409c318e518cb43" gracePeriod=30 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.145836 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.146215 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="4adf1e5a-318e-4ff3-a3e3-9af810860f12" containerName="nova-cell1-conductor-conductor" containerID="cri-o://51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9" gracePeriod=30 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.172999 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.173291 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-log" containerID="cri-o://18f4a09d21c8d0312c61833161ae0e52a65bfeb581a657ba4050c6dba468083b" gracePeriod=30 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.173872 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-metadata" containerID="cri-o://8cf1b4a2a70c8036c7b4eda3f909c4d2fadfe89bf04ba4ef470466c867cc3f95" gracePeriod=30 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.185021 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.185354 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1099964c-7106-4254-9bd5-551e2c73f1b8" containerName="nova-scheduler-scheduler" containerID="cri-o://5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38" gracePeriod=30 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.212273 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.212460 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e13167b4-439a-4071-af11-a524cd4d214e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://c662db2f16b33fb17267eee2d4d3f7d0dbece41314464b19df007b01164d6f7d" gracePeriod=30 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.306427 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.306628 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="dac90ac1-ccec-4016-be71-26d09ca77ae9" containerName="nova-cell0-conductor-conductor" containerID="cri-o://ce29af541c38e6ae0a8c967e0b55f541ed938943b2097cbb748168b7b3b01aaa" gracePeriod=30 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.346875 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" path="/var/lib/kubelet/pods/eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1/volumes" Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.443491 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="e13167b4-439a-4071-af11-a524cd4d214e" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.68:6080/vnc_lite.html\": dial tcp 10.217.1.68:6080: connect: connection refused" Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.903404 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"affe754f-50b9-4d77-9df2-027670584ddf","Type":"ContainerDied","Data":"6393ac159e85a10e047d0ef5367d40a9156fd4600cf811b07ff519153ce6d797"} Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.903996 4733 generic.go:334] "Generic (PLEG): container finished" podID="affe754f-50b9-4d77-9df2-027670584ddf" containerID="6393ac159e85a10e047d0ef5367d40a9156fd4600cf811b07ff519153ce6d797" exitCode=143 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.908445 4733 generic.go:334] "Generic (PLEG): container finished" podID="34b65d76-f6f6-4840-961a-e0878c140d78" containerID="18f4a09d21c8d0312c61833161ae0e52a65bfeb581a657ba4050c6dba468083b" exitCode=143 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.908570 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b65d76-f6f6-4840-961a-e0878c140d78","Type":"ContainerDied","Data":"18f4a09d21c8d0312c61833161ae0e52a65bfeb581a657ba4050c6dba468083b"} Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.910741 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"464f9982-87db-4cc0-991d-83eb0148e61b","Type":"ContainerStarted","Data":"5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1"} Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.910855 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.917965 4733 generic.go:334] "Generic (PLEG): container finished" podID="e13167b4-439a-4071-af11-a524cd4d214e" containerID="c662db2f16b33fb17267eee2d4d3f7d0dbece41314464b19df007b01164d6f7d" exitCode=0 Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.918032 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e13167b4-439a-4071-af11-a524cd4d214e","Type":"ContainerDied","Data":"c662db2f16b33fb17267eee2d4d3f7d0dbece41314464b19df007b01164d6f7d"} Dec 04 19:17:16 crc kubenswrapper[4733]: I1204 19:17:16.943157 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.943132354 podStartE2EDuration="3.943132354s" podCreationTimestamp="2025-12-04 19:17:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:16.929045973 +0000 UTC m=+5898.884407019" watchObservedRunningTime="2025-12-04 19:17:16.943132354 +0000 UTC m=+5898.898493400" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.144094 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.280821 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2w7z\" (UniqueName: \"kubernetes.io/projected/e13167b4-439a-4071-af11-a524cd4d214e-kube-api-access-m2w7z\") pod \"e13167b4-439a-4071-af11-a524cd4d214e\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.280890 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-config-data\") pod \"e13167b4-439a-4071-af11-a524cd4d214e\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.281011 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-combined-ca-bundle\") pod \"e13167b4-439a-4071-af11-a524cd4d214e\" (UID: \"e13167b4-439a-4071-af11-a524cd4d214e\") " Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.288473 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.290387 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e13167b4-439a-4071-af11-a524cd4d214e-kube-api-access-m2w7z" (OuterVolumeSpecName: "kube-api-access-m2w7z") pod "e13167b4-439a-4071-af11-a524cd4d214e" (UID: "e13167b4-439a-4071-af11-a524cd4d214e"). InnerVolumeSpecName "kube-api-access-m2w7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.317618 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-config-data" (OuterVolumeSpecName: "config-data") pod "e13167b4-439a-4071-af11-a524cd4d214e" (UID: "e13167b4-439a-4071-af11-a524cd4d214e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.317877 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e13167b4-439a-4071-af11-a524cd4d214e" (UID: "e13167b4-439a-4071-af11-a524cd4d214e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.383812 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2w7z\" (UniqueName: \"kubernetes.io/projected/e13167b4-439a-4071-af11-a524cd4d214e-kube-api-access-m2w7z\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.383863 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.383876 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13167b4-439a-4071-af11-a524cd4d214e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.484647 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-config-data\") pod \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.485084 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc7kr\" (UniqueName: \"kubernetes.io/projected/4adf1e5a-318e-4ff3-a3e3-9af810860f12-kube-api-access-mc7kr\") pod \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.485288 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-combined-ca-bundle\") pod \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\" (UID: \"4adf1e5a-318e-4ff3-a3e3-9af810860f12\") " Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.488333 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4adf1e5a-318e-4ff3-a3e3-9af810860f12-kube-api-access-mc7kr" (OuterVolumeSpecName: "kube-api-access-mc7kr") pod "4adf1e5a-318e-4ff3-a3e3-9af810860f12" (UID: "4adf1e5a-318e-4ff3-a3e3-9af810860f12"). InnerVolumeSpecName "kube-api-access-mc7kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.510262 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-config-data" (OuterVolumeSpecName: "config-data") pod "4adf1e5a-318e-4ff3-a3e3-9af810860f12" (UID: "4adf1e5a-318e-4ff3-a3e3-9af810860f12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.517526 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4adf1e5a-318e-4ff3-a3e3-9af810860f12" (UID: "4adf1e5a-318e-4ff3-a3e3-9af810860f12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.586739 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc7kr\" (UniqueName: \"kubernetes.io/projected/4adf1e5a-318e-4ff3-a3e3-9af810860f12-kube-api-access-mc7kr\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.586775 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.586785 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adf1e5a-318e-4ff3-a3e3-9af810860f12-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:17 crc kubenswrapper[4733]: E1204 19:17:17.699298 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 19:17:17 crc kubenswrapper[4733]: E1204 19:17:17.701430 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 19:17:17 crc kubenswrapper[4733]: E1204 19:17:17.702944 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 19:17:17 crc kubenswrapper[4733]: E1204 19:17:17.703004 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1099964c-7106-4254-9bd5-551e2c73f1b8" containerName="nova-scheduler-scheduler" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.930394 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e13167b4-439a-4071-af11-a524cd4d214e","Type":"ContainerDied","Data":"af87d189a952302a7c4aa69ad6641164bf5176fc1d1d143bdf0ec345e97f3828"} Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.930501 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.930516 4733 scope.go:117] "RemoveContainer" containerID="c662db2f16b33fb17267eee2d4d3f7d0dbece41314464b19df007b01164d6f7d" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.933872 4733 generic.go:334] "Generic (PLEG): container finished" podID="4adf1e5a-318e-4ff3-a3e3-9af810860f12" containerID="51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9" exitCode=0 Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.933931 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.933936 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4adf1e5a-318e-4ff3-a3e3-9af810860f12","Type":"ContainerDied","Data":"51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9"} Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.934012 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4adf1e5a-318e-4ff3-a3e3-9af810860f12","Type":"ContainerDied","Data":"387a2e9076b4ffceeb4bd1ec1017ebf15872efc3918fe49be5454b504c74a444"} Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.954419 4733 scope.go:117] "RemoveContainer" containerID="51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.986816 4733 scope.go:117] "RemoveContainer" containerID="51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.986938 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:17:17 crc kubenswrapper[4733]: E1204 19:17:17.987571 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9\": container with ID starting with 51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9 not found: ID does not exist" containerID="51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9" Dec 04 19:17:17 crc kubenswrapper[4733]: I1204 19:17:17.987597 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9"} err="failed to get container status \"51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9\": rpc error: code = NotFound desc = could not find container \"51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9\": container with ID starting with 51db7fac19e47bbf78f10b6cd6f04f9fda9bfc63f1a658c899cfba708ae71ac9 not found: ID does not exist" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.006433 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.014999 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.023472 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.032897 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:17:18 crc kubenswrapper[4733]: E1204 19:17:18.033411 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13167b4-439a-4071-af11-a524cd4d214e" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.033436 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13167b4-439a-4071-af11-a524cd4d214e" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 19:17:18 crc kubenswrapper[4733]: E1204 19:17:18.033460 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerName="extract-content" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.033468 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerName="extract-content" Dec 04 19:17:18 crc kubenswrapper[4733]: E1204 19:17:18.033489 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerName="extract-utilities" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.033497 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerName="extract-utilities" Dec 04 19:17:18 crc kubenswrapper[4733]: E1204 19:17:18.033518 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerName="registry-server" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.033527 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerName="registry-server" Dec 04 19:17:18 crc kubenswrapper[4733]: E1204 19:17:18.033545 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4adf1e5a-318e-4ff3-a3e3-9af810860f12" containerName="nova-cell1-conductor-conductor" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.033554 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4adf1e5a-318e-4ff3-a3e3-9af810860f12" containerName="nova-cell1-conductor-conductor" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.033751 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13167b4-439a-4071-af11-a524cd4d214e" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.033772 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="eed3d00a-7daa-40a4-bdfa-5b1ec6fe32e1" containerName="registry-server" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.033785 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4adf1e5a-318e-4ff3-a3e3-9af810860f12" containerName="nova-cell1-conductor-conductor" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.034508 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.038758 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.043528 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.054240 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.055600 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.057619 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.062744 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.094924 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g76fr\" (UniqueName: \"kubernetes.io/projected/c2d32ba0-df74-456e-a81b-784712deb2eb-kube-api-access-g76fr\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.094976 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.094999 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.095090 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct8g7\" (UniqueName: \"kubernetes.io/projected/bfa626a2-12b3-4774-b1f5-404753b13f2e-kube-api-access-ct8g7\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.095115 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa626a2-12b3-4774-b1f5-404753b13f2e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.095151 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa626a2-12b3-4774-b1f5-404753b13f2e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.197284 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct8g7\" (UniqueName: \"kubernetes.io/projected/bfa626a2-12b3-4774-b1f5-404753b13f2e-kube-api-access-ct8g7\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.197661 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa626a2-12b3-4774-b1f5-404753b13f2e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.197829 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa626a2-12b3-4774-b1f5-404753b13f2e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.198005 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g76fr\" (UniqueName: \"kubernetes.io/projected/c2d32ba0-df74-456e-a81b-784712deb2eb-kube-api-access-g76fr\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.198112 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.198226 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.201721 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.201819 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfa626a2-12b3-4774-b1f5-404753b13f2e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.201986 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa626a2-12b3-4774-b1f5-404753b13f2e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.202728 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.213487 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct8g7\" (UniqueName: \"kubernetes.io/projected/bfa626a2-12b3-4774-b1f5-404753b13f2e-kube-api-access-ct8g7\") pod \"nova-cell1-novncproxy-0\" (UID: \"bfa626a2-12b3-4774-b1f5-404753b13f2e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.213714 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g76fr\" (UniqueName: \"kubernetes.io/projected/c2d32ba0-df74-456e-a81b-784712deb2eb-kube-api-access-g76fr\") pod \"nova-cell1-conductor-0\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.359243 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4adf1e5a-318e-4ff3-a3e3-9af810860f12" path="/var/lib/kubelet/pods/4adf1e5a-318e-4ff3-a3e3-9af810860f12/volumes" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.360631 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e13167b4-439a-4071-af11-a524cd4d214e" path="/var/lib/kubelet/pods/e13167b4-439a-4071-af11-a524cd4d214e/volumes" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.375690 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.386419 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.923365 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.950934 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c2d32ba0-df74-456e-a81b-784712deb2eb","Type":"ContainerStarted","Data":"2b5c971cc5e33025ca24836b121c7295ae7509a03e1447c708bb3ba30276def4"} Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.957059 4733 generic.go:334] "Generic (PLEG): container finished" podID="dac90ac1-ccec-4016-be71-26d09ca77ae9" containerID="ce29af541c38e6ae0a8c967e0b55f541ed938943b2097cbb748168b7b3b01aaa" exitCode=0 Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.957117 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dac90ac1-ccec-4016-be71-26d09ca77ae9","Type":"ContainerDied","Data":"ce29af541c38e6ae0a8c967e0b55f541ed938943b2097cbb748168b7b3b01aaa"} Dec 04 19:17:18 crc kubenswrapper[4733]: I1204 19:17:18.995071 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 19:17:19 crc kubenswrapper[4733]: W1204 19:17:19.002105 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfa626a2_12b3_4774_b1f5_404753b13f2e.slice/crio-e0e728ee2e86f8d017208db5b946a9036e5b69f1c2e8b925de458ebc5ab34e35 WatchSource:0}: Error finding container e0e728ee2e86f8d017208db5b946a9036e5b69f1c2e8b925de458ebc5ab34e35: Status 404 returned error can't find the container with id e0e728ee2e86f8d017208db5b946a9036e5b69f1c2e8b925de458ebc5ab34e35 Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.039411 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.221384 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-config-data\") pod \"dac90ac1-ccec-4016-be71-26d09ca77ae9\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.221461 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4nk7\" (UniqueName: \"kubernetes.io/projected/dac90ac1-ccec-4016-be71-26d09ca77ae9-kube-api-access-m4nk7\") pod \"dac90ac1-ccec-4016-be71-26d09ca77ae9\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.221614 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-combined-ca-bundle\") pod \"dac90ac1-ccec-4016-be71-26d09ca77ae9\" (UID: \"dac90ac1-ccec-4016-be71-26d09ca77ae9\") " Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.227866 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dac90ac1-ccec-4016-be71-26d09ca77ae9-kube-api-access-m4nk7" (OuterVolumeSpecName: "kube-api-access-m4nk7") pod "dac90ac1-ccec-4016-be71-26d09ca77ae9" (UID: "dac90ac1-ccec-4016-be71-26d09ca77ae9"). InnerVolumeSpecName "kube-api-access-m4nk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.253408 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-config-data" (OuterVolumeSpecName: "config-data") pod "dac90ac1-ccec-4016-be71-26d09ca77ae9" (UID: "dac90ac1-ccec-4016-be71-26d09ca77ae9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.256165 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dac90ac1-ccec-4016-be71-26d09ca77ae9" (UID: "dac90ac1-ccec-4016-be71-26d09ca77ae9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.324186 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.324244 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4nk7\" (UniqueName: \"kubernetes.io/projected/dac90ac1-ccec-4016-be71-26d09ca77ae9-kube-api-access-m4nk7\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.324260 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac90ac1-ccec-4016-be71-26d09ca77ae9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.332662 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.77:8775/\": read tcp 10.217.0.2:49838->10.217.1.77:8775: read: connection reset by peer" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.332720 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.77:8775/\": read tcp 10.217.0.2:49846->10.217.1.77:8775: read: connection reset by peer" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.991846 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dac90ac1-ccec-4016-be71-26d09ca77ae9","Type":"ContainerDied","Data":"43ae9f0268adc4487c41224be38e23d8ab7e38bd8e74b16a6a777c28d75b46e2"} Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.991940 4733 scope.go:117] "RemoveContainer" containerID="ce29af541c38e6ae0a8c967e0b55f541ed938943b2097cbb748168b7b3b01aaa" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.991946 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.997698 4733 generic.go:334] "Generic (PLEG): container finished" podID="affe754f-50b9-4d77-9df2-027670584ddf" containerID="c13f6bfc3ac4e3925c706789f94e51fb144bf32d337a9c4f2409c318e518cb43" exitCode=0 Dec 04 19:17:19 crc kubenswrapper[4733]: I1204 19:17:19.997765 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"affe754f-50b9-4d77-9df2-027670584ddf","Type":"ContainerDied","Data":"c13f6bfc3ac4e3925c706789f94e51fb144bf32d337a9c4f2409c318e518cb43"} Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.001742 4733 generic.go:334] "Generic (PLEG): container finished" podID="34b65d76-f6f6-4840-961a-e0878c140d78" containerID="8cf1b4a2a70c8036c7b4eda3f909c4d2fadfe89bf04ba4ef470466c867cc3f95" exitCode=0 Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.001778 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b65d76-f6f6-4840-961a-e0878c140d78","Type":"ContainerDied","Data":"8cf1b4a2a70c8036c7b4eda3f909c4d2fadfe89bf04ba4ef470466c867cc3f95"} Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.007281 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bfa626a2-12b3-4774-b1f5-404753b13f2e","Type":"ContainerStarted","Data":"2d99a8fbb50da7c72b6ffbc03ac71e33e03762e7a2fc260e13417fa7f3d6126b"} Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.007320 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bfa626a2-12b3-4774-b1f5-404753b13f2e","Type":"ContainerStarted","Data":"e0e728ee2e86f8d017208db5b946a9036e5b69f1c2e8b925de458ebc5ab34e35"} Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.011880 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c2d32ba0-df74-456e-a81b-784712deb2eb","Type":"ContainerStarted","Data":"019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee"} Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.012755 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.032464 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.03244576 podStartE2EDuration="3.03244576s" podCreationTimestamp="2025-12-04 19:17:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:20.026940631 +0000 UTC m=+5901.982301697" watchObservedRunningTime="2025-12-04 19:17:20.03244576 +0000 UTC m=+5901.987806816" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.064296 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.064276171 podStartE2EDuration="3.064276171s" podCreationTimestamp="2025-12-04 19:17:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:20.044677171 +0000 UTC m=+5902.000038217" watchObservedRunningTime="2025-12-04 19:17:20.064276171 +0000 UTC m=+5902.019637217" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.081247 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.093139 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.102220 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:17:20 crc kubenswrapper[4733]: E1204 19:17:20.102581 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dac90ac1-ccec-4016-be71-26d09ca77ae9" containerName="nova-cell0-conductor-conductor" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.102598 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac90ac1-ccec-4016-be71-26d09ca77ae9" containerName="nova-cell0-conductor-conductor" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.102758 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="dac90ac1-ccec-4016-be71-26d09ca77ae9" containerName="nova-cell0-conductor-conductor" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.103372 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.106116 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.112075 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.238902 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.238951 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.238982 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xljmc\" (UniqueName: \"kubernetes.io/projected/ef61747b-f475-4934-974d-7cf53750963b-kube-api-access-xljmc\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.335700 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.340555 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.340606 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.340626 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xljmc\" (UniqueName: \"kubernetes.io/projected/ef61747b-f475-4934-974d-7cf53750963b-kube-api-access-xljmc\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.347688 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dac90ac1-ccec-4016-be71-26d09ca77ae9" path="/var/lib/kubelet/pods/dac90ac1-ccec-4016-be71-26d09ca77ae9/volumes" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.348661 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.366977 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.370371 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xljmc\" (UniqueName: \"kubernetes.io/projected/ef61747b-f475-4934-974d-7cf53750963b-kube-api-access-xljmc\") pod \"nova-cell0-conductor-0\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.407108 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.431634 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.441595 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-config-data\") pod \"affe754f-50b9-4d77-9df2-027670584ddf\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.441639 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrmc6\" (UniqueName: \"kubernetes.io/projected/affe754f-50b9-4d77-9df2-027670584ddf-kube-api-access-xrmc6\") pod \"affe754f-50b9-4d77-9df2-027670584ddf\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.441691 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-combined-ca-bundle\") pod \"affe754f-50b9-4d77-9df2-027670584ddf\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.441740 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affe754f-50b9-4d77-9df2-027670584ddf-logs\") pod \"affe754f-50b9-4d77-9df2-027670584ddf\" (UID: \"affe754f-50b9-4d77-9df2-027670584ddf\") " Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.447992 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/affe754f-50b9-4d77-9df2-027670584ddf-logs" (OuterVolumeSpecName: "logs") pod "affe754f-50b9-4d77-9df2-027670584ddf" (UID: "affe754f-50b9-4d77-9df2-027670584ddf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.452438 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/affe754f-50b9-4d77-9df2-027670584ddf-kube-api-access-xrmc6" (OuterVolumeSpecName: "kube-api-access-xrmc6") pod "affe754f-50b9-4d77-9df2-027670584ddf" (UID: "affe754f-50b9-4d77-9df2-027670584ddf"). InnerVolumeSpecName "kube-api-access-xrmc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.482055 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-config-data" (OuterVolumeSpecName: "config-data") pod "affe754f-50b9-4d77-9df2-027670584ddf" (UID: "affe754f-50b9-4d77-9df2-027670584ddf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.482155 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "affe754f-50b9-4d77-9df2-027670584ddf" (UID: "affe754f-50b9-4d77-9df2-027670584ddf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.543583 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b65d76-f6f6-4840-961a-e0878c140d78-logs\") pod \"34b65d76-f6f6-4840-961a-e0878c140d78\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.543923 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-config-data\") pod \"34b65d76-f6f6-4840-961a-e0878c140d78\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.544132 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-combined-ca-bundle\") pod \"34b65d76-f6f6-4840-961a-e0878c140d78\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.544182 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqh7l\" (UniqueName: \"kubernetes.io/projected/34b65d76-f6f6-4840-961a-e0878c140d78-kube-api-access-nqh7l\") pod \"34b65d76-f6f6-4840-961a-e0878c140d78\" (UID: \"34b65d76-f6f6-4840-961a-e0878c140d78\") " Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.545562 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34b65d76-f6f6-4840-961a-e0878c140d78-logs" (OuterVolumeSpecName: "logs") pod "34b65d76-f6f6-4840-961a-e0878c140d78" (UID: "34b65d76-f6f6-4840-961a-e0878c140d78"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.547416 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affe754f-50b9-4d77-9df2-027670584ddf-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.547437 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b65d76-f6f6-4840-961a-e0878c140d78-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.547449 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.547462 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrmc6\" (UniqueName: \"kubernetes.io/projected/affe754f-50b9-4d77-9df2-027670584ddf-kube-api-access-xrmc6\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.547474 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affe754f-50b9-4d77-9df2-027670584ddf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.555688 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34b65d76-f6f6-4840-961a-e0878c140d78-kube-api-access-nqh7l" (OuterVolumeSpecName: "kube-api-access-nqh7l") pod "34b65d76-f6f6-4840-961a-e0878c140d78" (UID: "34b65d76-f6f6-4840-961a-e0878c140d78"). InnerVolumeSpecName "kube-api-access-nqh7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.576718 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-config-data" (OuterVolumeSpecName: "config-data") pod "34b65d76-f6f6-4840-961a-e0878c140d78" (UID: "34b65d76-f6f6-4840-961a-e0878c140d78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.596905 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34b65d76-f6f6-4840-961a-e0878c140d78" (UID: "34b65d76-f6f6-4840-961a-e0878c140d78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.649465 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.649498 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqh7l\" (UniqueName: \"kubernetes.io/projected/34b65d76-f6f6-4840-961a-e0878c140d78-kube-api-access-nqh7l\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.649509 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b65d76-f6f6-4840-961a-e0878c140d78-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:20 crc kubenswrapper[4733]: I1204 19:17:20.910228 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 19:17:20 crc kubenswrapper[4733]: W1204 19:17:20.914931 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef61747b_f475_4934_974d_7cf53750963b.slice/crio-6210138a8804afba410828ee650a6b740ac3fe0971496c2e8f3c23d07fbe3ae5 WatchSource:0}: Error finding container 6210138a8804afba410828ee650a6b740ac3fe0971496c2e8f3c23d07fbe3ae5: Status 404 returned error can't find the container with id 6210138a8804afba410828ee650a6b740ac3fe0971496c2e8f3c23d07fbe3ae5 Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.026701 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b65d76-f6f6-4840-961a-e0878c140d78","Type":"ContainerDied","Data":"68e8a0517089fe7df20037e8ad4ab18fb28834ce255743080f1fb203d3695fcd"} Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.028457 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ef61747b-f475-4934-974d-7cf53750963b","Type":"ContainerStarted","Data":"6210138a8804afba410828ee650a6b740ac3fe0971496c2e8f3c23d07fbe3ae5"} Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.026767 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.028497 4733 scope.go:117] "RemoveContainer" containerID="8cf1b4a2a70c8036c7b4eda3f909c4d2fadfe89bf04ba4ef470466c867cc3f95" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.036459 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.036454 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"affe754f-50b9-4d77-9df2-027670584ddf","Type":"ContainerDied","Data":"0310c93c1ce1267fcb27f958c805506cea0240d111bca49804923b712abefd1b"} Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.069196 4733 scope.go:117] "RemoveContainer" containerID="18f4a09d21c8d0312c61833161ae0e52a65bfeb581a657ba4050c6dba468083b" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.110470 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.122612 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.130450 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.131240 4733 scope.go:117] "RemoveContainer" containerID="c13f6bfc3ac4e3925c706789f94e51fb144bf32d337a9c4f2409c318e518cb43" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.145375 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164119 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 19:17:21 crc kubenswrapper[4733]: E1204 19:17:21.164507 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-log" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164524 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-log" Dec 04 19:17:21 crc kubenswrapper[4733]: E1204 19:17:21.164538 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-log" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164546 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-log" Dec 04 19:17:21 crc kubenswrapper[4733]: E1204 19:17:21.164562 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-metadata" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164568 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-metadata" Dec 04 19:17:21 crc kubenswrapper[4733]: E1204 19:17:21.164592 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-api" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164598 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-api" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164776 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-log" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164803 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-log" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164820 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="affe754f-50b9-4d77-9df2-027670584ddf" containerName="nova-api-api" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.164838 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" containerName="nova-metadata-metadata" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.165859 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.168256 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.174511 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-logs\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.176697 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.177778 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-config-data\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.178183 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-444dk\" (UniqueName: \"kubernetes.io/projected/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-kube-api-access-444dk\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.198581 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.211075 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.212454 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.213663 4733 scope.go:117] "RemoveContainer" containerID="6393ac159e85a10e047d0ef5367d40a9156fd4600cf811b07ff519153ce6d797" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.215123 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.220022 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.279221 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.279280 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-config-data\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.279324 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-444dk\" (UniqueName: \"kubernetes.io/projected/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-kube-api-access-444dk\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.279352 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-config-data\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.279393 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-logs\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.279429 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.279491 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b248d65a-3d4e-4c73-a49c-53864042c710-logs\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.279523 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2955\" (UniqueName: \"kubernetes.io/projected/b248d65a-3d4e-4c73-a49c-53864042c710-kube-api-access-r2955\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.280030 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-logs\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.285506 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.285527 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-config-data\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.295283 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-444dk\" (UniqueName: \"kubernetes.io/projected/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-kube-api-access-444dk\") pod \"nova-api-0\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.381518 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-config-data\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.381599 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.381670 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b248d65a-3d4e-4c73-a49c-53864042c710-logs\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.381728 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2955\" (UniqueName: \"kubernetes.io/projected/b248d65a-3d4e-4c73-a49c-53864042c710-kube-api-access-r2955\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.382759 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b248d65a-3d4e-4c73-a49c-53864042c710-logs\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.394704 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.395051 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-config-data\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.397649 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2955\" (UniqueName: \"kubernetes.io/projected/b248d65a-3d4e-4c73-a49c-53864042c710-kube-api-access-r2955\") pod \"nova-metadata-0\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.498627 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.543244 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.724910 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.788125 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-config-data\") pod \"1099964c-7106-4254-9bd5-551e2c73f1b8\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.788434 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-combined-ca-bundle\") pod \"1099964c-7106-4254-9bd5-551e2c73f1b8\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.788586 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw8bl\" (UniqueName: \"kubernetes.io/projected/1099964c-7106-4254-9bd5-551e2c73f1b8-kube-api-access-bw8bl\") pod \"1099964c-7106-4254-9bd5-551e2c73f1b8\" (UID: \"1099964c-7106-4254-9bd5-551e2c73f1b8\") " Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.794360 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1099964c-7106-4254-9bd5-551e2c73f1b8-kube-api-access-bw8bl" (OuterVolumeSpecName: "kube-api-access-bw8bl") pod "1099964c-7106-4254-9bd5-551e2c73f1b8" (UID: "1099964c-7106-4254-9bd5-551e2c73f1b8"). InnerVolumeSpecName "kube-api-access-bw8bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.819731 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-config-data" (OuterVolumeSpecName: "config-data") pod "1099964c-7106-4254-9bd5-551e2c73f1b8" (UID: "1099964c-7106-4254-9bd5-551e2c73f1b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.820300 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1099964c-7106-4254-9bd5-551e2c73f1b8" (UID: "1099964c-7106-4254-9bd5-551e2c73f1b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.890175 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.890212 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1099964c-7106-4254-9bd5-551e2c73f1b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:21 crc kubenswrapper[4733]: I1204 19:17:21.890244 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw8bl\" (UniqueName: \"kubernetes.io/projected/1099964c-7106-4254-9bd5-551e2c73f1b8-kube-api-access-bw8bl\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.005104 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.046866 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60","Type":"ContainerStarted","Data":"f663f4c4e891851f9cbead8ca9b169acdd8974b27e7b74ead0ca61c40fa952bb"} Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.049053 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ef61747b-f475-4934-974d-7cf53750963b","Type":"ContainerStarted","Data":"f47f2b93431904bee107f24ffa8f58bf14b54ed1f3f0ce1b95d4d0d724abc65d"} Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.050279 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.052256 4733 generic.go:334] "Generic (PLEG): container finished" podID="1099964c-7106-4254-9bd5-551e2c73f1b8" containerID="5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38" exitCode=0 Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.052298 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1099964c-7106-4254-9bd5-551e2c73f1b8","Type":"ContainerDied","Data":"5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38"} Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.052316 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1099964c-7106-4254-9bd5-551e2c73f1b8","Type":"ContainerDied","Data":"303e67224f294ebbd3dad9a9622475bd6b7d88916d61011e5f7d80f8596a677d"} Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.052332 4733 scope.go:117] "RemoveContainer" containerID="5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.052424 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.082244 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.08222403 podStartE2EDuration="2.08222403s" podCreationTimestamp="2025-12-04 19:17:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:22.076324951 +0000 UTC m=+5904.031685987" watchObservedRunningTime="2025-12-04 19:17:22.08222403 +0000 UTC m=+5904.037585076" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.083719 4733 scope.go:117] "RemoveContainer" containerID="5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38" Dec 04 19:17:22 crc kubenswrapper[4733]: E1204 19:17:22.084639 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38\": container with ID starting with 5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38 not found: ID does not exist" containerID="5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.084699 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38"} err="failed to get container status \"5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38\": rpc error: code = NotFound desc = could not find container \"5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38\": container with ID starting with 5a6cefaa54ac3c6b1d7df529b156a67b9365c0cdae52ae4f635e8160c1f14a38 not found: ID does not exist" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.124997 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.141490 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.157595 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.168405 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:17:22 crc kubenswrapper[4733]: E1204 19:17:22.169081 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1099964c-7106-4254-9bd5-551e2c73f1b8" containerName="nova-scheduler-scheduler" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.169106 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1099964c-7106-4254-9bd5-551e2c73f1b8" containerName="nova-scheduler-scheduler" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.169347 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1099964c-7106-4254-9bd5-551e2c73f1b8" containerName="nova-scheduler-scheduler" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.170177 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.172496 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.181845 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.296750 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-config-data\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.297101 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wsvh\" (UniqueName: \"kubernetes.io/projected/fffc6200-ffec-4120-9e25-feb4982180c3-kube-api-access-4wsvh\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.297149 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.352400 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1099964c-7106-4254-9bd5-551e2c73f1b8" path="/var/lib/kubelet/pods/1099964c-7106-4254-9bd5-551e2c73f1b8/volumes" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.353327 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34b65d76-f6f6-4840-961a-e0878c140d78" path="/var/lib/kubelet/pods/34b65d76-f6f6-4840-961a-e0878c140d78/volumes" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.355409 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="affe754f-50b9-4d77-9df2-027670584ddf" path="/var/lib/kubelet/pods/affe754f-50b9-4d77-9df2-027670584ddf/volumes" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.398642 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-config-data\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.398688 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wsvh\" (UniqueName: \"kubernetes.io/projected/fffc6200-ffec-4120-9e25-feb4982180c3-kube-api-access-4wsvh\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.398736 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.402572 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.402753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-config-data\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.421522 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wsvh\" (UniqueName: \"kubernetes.io/projected/fffc6200-ffec-4120-9e25-feb4982180c3-kube-api-access-4wsvh\") pod \"nova-scheduler-0\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " pod="openstack/nova-scheduler-0" Dec 04 19:17:22 crc kubenswrapper[4733]: I1204 19:17:22.570305 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.019035 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.077941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b248d65a-3d4e-4c73-a49c-53864042c710","Type":"ContainerStarted","Data":"5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876"} Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.078004 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b248d65a-3d4e-4c73-a49c-53864042c710","Type":"ContainerStarted","Data":"6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205"} Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.078022 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b248d65a-3d4e-4c73-a49c-53864042c710","Type":"ContainerStarted","Data":"37bfcf2b5a50ae75b10e7a992aea2892d9a04ba6b3c2b332de157eb88247f666"} Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.082412 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fffc6200-ffec-4120-9e25-feb4982180c3","Type":"ContainerStarted","Data":"99af1f1058b24405ec7e48d9c9fcfb7105d773e0ed03aa522eb846075f54aa83"} Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.085454 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60","Type":"ContainerStarted","Data":"e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e"} Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.085501 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60","Type":"ContainerStarted","Data":"6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149"} Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.110181 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.110160473 podStartE2EDuration="2.110160473s" podCreationTimestamp="2025-12-04 19:17:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:23.105825745 +0000 UTC m=+5905.061186801" watchObservedRunningTime="2025-12-04 19:17:23.110160473 +0000 UTC m=+5905.065521529" Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.142744 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.142726103 podStartE2EDuration="2.142726103s" podCreationTimestamp="2025-12-04 19:17:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:23.134386637 +0000 UTC m=+5905.089747683" watchObservedRunningTime="2025-12-04 19:17:23.142726103 +0000 UTC m=+5905.098087149" Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.386858 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.480000 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.591836 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8675c4d89f-t6wg9"] Dec 04 19:17:23 crc kubenswrapper[4733]: I1204 19:17:23.592201 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" podUID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerName="dnsmasq-dns" containerID="cri-o://471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d" gracePeriod=10 Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.082997 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.093192 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fffc6200-ffec-4120-9e25-feb4982180c3","Type":"ContainerStarted","Data":"0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2"} Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.095191 4733 generic.go:334] "Generic (PLEG): container finished" podID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerID="471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d" exitCode=0 Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.095285 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.095665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" event={"ID":"cd9ddcea-451e-4d0b-94f4-821c3765eb0a","Type":"ContainerDied","Data":"471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d"} Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.095695 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" event={"ID":"cd9ddcea-451e-4d0b-94f4-821c3765eb0a","Type":"ContainerDied","Data":"f509e673ed3f416ab37c774ff33a3ac1782949925f9c477285ca84b6ec39fdcb"} Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.095710 4733 scope.go:117] "RemoveContainer" containerID="471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.136221 4733 scope.go:117] "RemoveContainer" containerID="c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.136738 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.136719938 podStartE2EDuration="2.136719938s" podCreationTimestamp="2025-12-04 19:17:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:24.128581047 +0000 UTC m=+5906.083942103" watchObservedRunningTime="2025-12-04 19:17:24.136719938 +0000 UTC m=+5906.092080974" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.146559 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48bfc\" (UniqueName: \"kubernetes.io/projected/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-kube-api-access-48bfc\") pod \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.146627 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-sb\") pod \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.146701 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-dns-svc\") pod \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.146881 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-config\") pod \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.146929 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-nb\") pod \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\" (UID: \"cd9ddcea-451e-4d0b-94f4-821c3765eb0a\") " Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.153589 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-kube-api-access-48bfc" (OuterVolumeSpecName: "kube-api-access-48bfc") pod "cd9ddcea-451e-4d0b-94f4-821c3765eb0a" (UID: "cd9ddcea-451e-4d0b-94f4-821c3765eb0a"). InnerVolumeSpecName "kube-api-access-48bfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.164378 4733 scope.go:117] "RemoveContainer" containerID="471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d" Dec 04 19:17:24 crc kubenswrapper[4733]: E1204 19:17:24.165014 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d\": container with ID starting with 471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d not found: ID does not exist" containerID="471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.165131 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d"} err="failed to get container status \"471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d\": rpc error: code = NotFound desc = could not find container \"471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d\": container with ID starting with 471f43ee95d155dad4dfac73c479760e74f55a60e980ce3741582975cd600e5d not found: ID does not exist" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.165257 4733 scope.go:117] "RemoveContainer" containerID="c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2" Dec 04 19:17:24 crc kubenswrapper[4733]: E1204 19:17:24.165567 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2\": container with ID starting with c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2 not found: ID does not exist" containerID="c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.165676 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2"} err="failed to get container status \"c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2\": rpc error: code = NotFound desc = could not find container \"c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2\": container with ID starting with c6b234ca53aa9f325363492376b12c64d154e926f6e76423e9904f0f0a0369d2 not found: ID does not exist" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.194155 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-config" (OuterVolumeSpecName: "config") pod "cd9ddcea-451e-4d0b-94f4-821c3765eb0a" (UID: "cd9ddcea-451e-4d0b-94f4-821c3765eb0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.198665 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cd9ddcea-451e-4d0b-94f4-821c3765eb0a" (UID: "cd9ddcea-451e-4d0b-94f4-821c3765eb0a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.199266 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cd9ddcea-451e-4d0b-94f4-821c3765eb0a" (UID: "cd9ddcea-451e-4d0b-94f4-821c3765eb0a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.204366 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cd9ddcea-451e-4d0b-94f4-821c3765eb0a" (UID: "cd9ddcea-451e-4d0b-94f4-821c3765eb0a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.250550 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.250650 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.250674 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48bfc\" (UniqueName: \"kubernetes.io/projected/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-kube-api-access-48bfc\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.250718 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.250732 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd9ddcea-451e-4d0b-94f4-821c3765eb0a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.437323 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8675c4d89f-t6wg9"] Dec 04 19:17:24 crc kubenswrapper[4733]: I1204 19:17:24.444935 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8675c4d89f-t6wg9"] Dec 04 19:17:25 crc kubenswrapper[4733]: I1204 19:17:25.806518 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 19:17:26 crc kubenswrapper[4733]: I1204 19:17:26.348814 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" path="/var/lib/kubelet/pods/cd9ddcea-451e-4d0b-94f4-821c3765eb0a/volumes" Dec 04 19:17:26 crc kubenswrapper[4733]: I1204 19:17:26.544090 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 19:17:26 crc kubenswrapper[4733]: I1204 19:17:26.544892 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 19:17:27 crc kubenswrapper[4733]: I1204 19:17:27.336246 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:17:27 crc kubenswrapper[4733]: E1204 19:17:27.336706 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:17:27 crc kubenswrapper[4733]: I1204 19:17:27.570466 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 19:17:28 crc kubenswrapper[4733]: I1204 19:17:28.387537 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:28 crc kubenswrapper[4733]: I1204 19:17:28.404924 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:28 crc kubenswrapper[4733]: I1204 19:17:28.416115 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 19:17:29 crc kubenswrapper[4733]: I1204 19:17:29.025387 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8675c4d89f-t6wg9" podUID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.80:5353: i/o timeout" Dec 04 19:17:29 crc kubenswrapper[4733]: I1204 19:17:29.184111 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 19:17:30 crc kubenswrapper[4733]: I1204 19:17:30.461235 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 19:17:31 crc kubenswrapper[4733]: I1204 19:17:31.498845 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 19:17:31 crc kubenswrapper[4733]: I1204 19:17:31.499141 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 19:17:31 crc kubenswrapper[4733]: I1204 19:17:31.543666 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 19:17:31 crc kubenswrapper[4733]: I1204 19:17:31.543707 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 19:17:32 crc kubenswrapper[4733]: I1204 19:17:32.540070 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:17:32 crc kubenswrapper[4733]: I1204 19:17:32.540257 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:17:32 crc kubenswrapper[4733]: I1204 19:17:32.571329 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 19:17:32 crc kubenswrapper[4733]: I1204 19:17:32.603453 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 19:17:32 crc kubenswrapper[4733]: I1204 19:17:32.627022 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.91:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:17:32 crc kubenswrapper[4733]: I1204 19:17:32.627367 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.91:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 19:17:33 crc kubenswrapper[4733]: I1204 19:17:33.236679 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 19:17:39 crc kubenswrapper[4733]: I1204 19:17:39.336035 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:17:39 crc kubenswrapper[4733]: E1204 19:17:39.336862 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.506266 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.506874 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.507227 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.507244 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.511538 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.511708 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.553440 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.561424 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 19:17:41 crc kubenswrapper[4733]: I1204 19:17:41.562680 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 19:17:42 crc kubenswrapper[4733]: I1204 19:17:42.319427 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.178657 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:43 crc kubenswrapper[4733]: E1204 19:17:43.179345 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerName="init" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.179360 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerName="init" Dec 04 19:17:43 crc kubenswrapper[4733]: E1204 19:17:43.179393 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerName="dnsmasq-dns" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.179399 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerName="dnsmasq-dns" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.179583 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd9ddcea-451e-4d0b-94f4-821c3765eb0a" containerName="dnsmasq-dns" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.180612 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.182727 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.190244 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.343049 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-scripts\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.343102 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60698b9c-8516-4023-a27c-e6bcaad6ee7a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.343196 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.343226 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.343258 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2rjq\" (UniqueName: \"kubernetes.io/projected/60698b9c-8516-4023-a27c-e6bcaad6ee7a-kube-api-access-m2rjq\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.343321 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.444596 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.444775 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-scripts\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.444830 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60698b9c-8516-4023-a27c-e6bcaad6ee7a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.444893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.444926 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.444956 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2rjq\" (UniqueName: \"kubernetes.io/projected/60698b9c-8516-4023-a27c-e6bcaad6ee7a-kube-api-access-m2rjq\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.445235 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60698b9c-8516-4023-a27c-e6bcaad6ee7a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.451055 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.455289 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-scripts\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.455775 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.457228 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.468518 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2rjq\" (UniqueName: \"kubernetes.io/projected/60698b9c-8516-4023-a27c-e6bcaad6ee7a-kube-api-access-m2rjq\") pod \"cinder-scheduler-0\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.508147 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 19:17:43 crc kubenswrapper[4733]: I1204 19:17:43.960578 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:44 crc kubenswrapper[4733]: I1204 19:17:44.375967 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60698b9c-8516-4023-a27c-e6bcaad6ee7a","Type":"ContainerStarted","Data":"ec5434c1c3643a2c7cda864aa8ee6be7e9d9d3d7d80ca2b1b5bca0cf12e0c2a6"} Dec 04 19:17:44 crc kubenswrapper[4733]: I1204 19:17:44.487456 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:44 crc kubenswrapper[4733]: I1204 19:17:44.487735 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" containerName="cinder-api-log" containerID="cri-o://ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d" gracePeriod=30 Dec 04 19:17:44 crc kubenswrapper[4733]: I1204 19:17:44.487857 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" containerName="cinder-api" containerID="cri-o://5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1" gracePeriod=30 Dec 04 19:17:44 crc kubenswrapper[4733]: I1204 19:17:44.913065 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 04 19:17:44 crc kubenswrapper[4733]: I1204 19:17:44.915994 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:44 crc kubenswrapper[4733]: I1204 19:17:44.920627 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 04 19:17:44 crc kubenswrapper[4733]: I1204 19:17:44.926295 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078236 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-dev\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078276 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078300 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078336 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-sys\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078393 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078408 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078433 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bb94a69c-3457-498f-a83a-43e375f89dfd-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078447 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078589 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078638 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078679 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hswmw\" (UniqueName: \"kubernetes.io/projected/bb94a69c-3457-498f-a83a-43e375f89dfd-kube-api-access-hswmw\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078785 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-run\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078844 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078915 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.078976 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.079007 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180298 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180359 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180392 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180414 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180444 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-dev\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180459 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180507 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-sys\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180530 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180545 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180545 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-dev\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180571 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bb94a69c-3457-498f-a83a-43e375f89dfd-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180587 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180600 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180606 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180625 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180664 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180645 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180715 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180745 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180742 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hswmw\" (UniqueName: \"kubernetes.io/projected/bb94a69c-3457-498f-a83a-43e375f89dfd-kube-api-access-hswmw\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180813 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-run\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180828 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180851 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-sys\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180873 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-run\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.180933 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bb94a69c-3457-498f-a83a-43e375f89dfd-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.185128 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.185504 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.185606 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.186373 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb94a69c-3457-498f-a83a-43e375f89dfd-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.186406 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bb94a69c-3457-498f-a83a-43e375f89dfd-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.212364 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hswmw\" (UniqueName: \"kubernetes.io/projected/bb94a69c-3457-498f-a83a-43e375f89dfd-kube-api-access-hswmw\") pod \"cinder-volume-volume1-0\" (UID: \"bb94a69c-3457-498f-a83a-43e375f89dfd\") " pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.238082 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.362361 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60698b9c-8516-4023-a27c-e6bcaad6ee7a","Type":"ContainerStarted","Data":"44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23"} Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.362637 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60698b9c-8516-4023-a27c-e6bcaad6ee7a","Type":"ContainerStarted","Data":"28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87"} Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.365474 4733 generic.go:334] "Generic (PLEG): container finished" podID="464f9982-87db-4cc0-991d-83eb0148e61b" containerID="ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d" exitCode=143 Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.365526 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"464f9982-87db-4cc0-991d-83eb0148e61b","Type":"ContainerDied","Data":"ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d"} Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.400699 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.400678179 podStartE2EDuration="2.400678179s" podCreationTimestamp="2025-12-04 19:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:45.386820065 +0000 UTC m=+5927.342181121" watchObservedRunningTime="2025-12-04 19:17:45.400678179 +0000 UTC m=+5927.356039215" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.473875 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.475309 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.480052 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.509934 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.595736 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-lib-modules\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.595813 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.595839 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.595864 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-dev\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.595883 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.595927 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-nvme\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.595943 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/639c8ab0-c11a-4cf1-b944-9caa6b91604e-ceph\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.595969 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-sys\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.596018 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-config-data-custom\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.596039 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-run\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.596055 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.596074 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-config-data\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.596093 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxt76\" (UniqueName: \"kubernetes.io/projected/639c8ab0-c11a-4cf1-b944-9caa6b91604e-kube-api-access-hxt76\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.596110 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.596126 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.596144 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-scripts\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-sys\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-config-data-custom\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697794 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-run\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697823 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697843 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-config-data\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxt76\" (UniqueName: \"kubernetes.io/projected/639c8ab0-c11a-4cf1-b944-9caa6b91604e-kube-api-access-hxt76\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697878 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697896 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697916 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-scripts\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697941 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-lib-modules\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697962 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.697986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.698010 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-dev\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.698031 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.698071 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-nvme\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.698089 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/639c8ab0-c11a-4cf1-b944-9caa6b91604e-ceph\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.698666 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-sys\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.700937 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-dev\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.701026 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-lib-modules\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.701061 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.702951 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-nvme\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.703029 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.706268 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.706633 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.706672 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-run\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.706707 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/639c8ab0-c11a-4cf1-b944-9caa6b91604e-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.707911 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-config-data-custom\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.709367 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.712423 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-config-data\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.714538 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/639c8ab0-c11a-4cf1-b944-9caa6b91604e-scripts\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.718274 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/639c8ab0-c11a-4cf1-b944-9caa6b91604e-ceph\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.725946 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxt76\" (UniqueName: \"kubernetes.io/projected/639c8ab0-c11a-4cf1-b944-9caa6b91604e-kube-api-access-hxt76\") pod \"cinder-backup-0\" (UID: \"639c8ab0-c11a-4cf1-b944-9caa6b91604e\") " pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.839846 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 04 19:17:45 crc kubenswrapper[4733]: I1204 19:17:45.881586 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 04 19:17:46 crc kubenswrapper[4733]: I1204 19:17:46.376128 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"bb94a69c-3457-498f-a83a-43e375f89dfd","Type":"ContainerStarted","Data":"6571495cfdf888a8e04e8a3f66cc57d03c994b700048d572ee71a3e5bb5dac1c"} Dec 04 19:17:46 crc kubenswrapper[4733]: I1204 19:17:46.418962 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 04 19:17:46 crc kubenswrapper[4733]: W1204 19:17:46.419600 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod639c8ab0_c11a_4cf1_b944_9caa6b91604e.slice/crio-fdde53a9025be6118fbc2ada0f221c1d280acc5dea9785cce4bfbdee8ffd30cf WatchSource:0}: Error finding container fdde53a9025be6118fbc2ada0f221c1d280acc5dea9785cce4bfbdee8ffd30cf: Status 404 returned error can't find the container with id fdde53a9025be6118fbc2ada0f221c1d280acc5dea9785cce4bfbdee8ffd30cf Dec 04 19:17:47 crc kubenswrapper[4733]: I1204 19:17:47.385539 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"bb94a69c-3457-498f-a83a-43e375f89dfd","Type":"ContainerStarted","Data":"c31c4f26bfd0409a0c3bb360289aa7f6bf01bca22782bc0f85c45d5b22041aa6"} Dec 04 19:17:47 crc kubenswrapper[4733]: I1204 19:17:47.386035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"bb94a69c-3457-498f-a83a-43e375f89dfd","Type":"ContainerStarted","Data":"e29ceb6a21c90c10bccee6226fa28c5c57e3f45f541d942f5c7f7b11ae30b4fb"} Dec 04 19:17:47 crc kubenswrapper[4733]: I1204 19:17:47.388023 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"639c8ab0-c11a-4cf1-b944-9caa6b91604e","Type":"ContainerStarted","Data":"685386c198be54a3404c185fd602a9276f62ae7e29978deb120dc6a5d6b6382d"} Dec 04 19:17:47 crc kubenswrapper[4733]: I1204 19:17:47.388049 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"639c8ab0-c11a-4cf1-b944-9caa6b91604e","Type":"ContainerStarted","Data":"fdde53a9025be6118fbc2ada0f221c1d280acc5dea9785cce4bfbdee8ffd30cf"} Dec 04 19:17:47 crc kubenswrapper[4733]: I1204 19:17:47.418457 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.565495524 podStartE2EDuration="3.418436773s" podCreationTimestamp="2025-12-04 19:17:44 +0000 UTC" firstStartedPulling="2025-12-04 19:17:45.897068175 +0000 UTC m=+5927.852429241" lastFinishedPulling="2025-12-04 19:17:46.750009444 +0000 UTC m=+5928.705370490" observedRunningTime="2025-12-04 19:17:47.407516808 +0000 UTC m=+5929.362877844" watchObservedRunningTime="2025-12-04 19:17:47.418436773 +0000 UTC m=+5929.373797819" Dec 04 19:17:47 crc kubenswrapper[4733]: I1204 19:17:47.962425 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.151140 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data-custom\") pod \"464f9982-87db-4cc0-991d-83eb0148e61b\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.151465 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data\") pod \"464f9982-87db-4cc0-991d-83eb0148e61b\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.151515 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-combined-ca-bundle\") pod \"464f9982-87db-4cc0-991d-83eb0148e61b\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.151569 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-scripts\") pod \"464f9982-87db-4cc0-991d-83eb0148e61b\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.151596 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/464f9982-87db-4cc0-991d-83eb0148e61b-logs\") pod \"464f9982-87db-4cc0-991d-83eb0148e61b\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.151689 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/464f9982-87db-4cc0-991d-83eb0148e61b-etc-machine-id\") pod \"464f9982-87db-4cc0-991d-83eb0148e61b\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.151769 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk5h4\" (UniqueName: \"kubernetes.io/projected/464f9982-87db-4cc0-991d-83eb0148e61b-kube-api-access-vk5h4\") pod \"464f9982-87db-4cc0-991d-83eb0148e61b\" (UID: \"464f9982-87db-4cc0-991d-83eb0148e61b\") " Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.152889 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/464f9982-87db-4cc0-991d-83eb0148e61b-logs" (OuterVolumeSpecName: "logs") pod "464f9982-87db-4cc0-991d-83eb0148e61b" (UID: "464f9982-87db-4cc0-991d-83eb0148e61b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.152950 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/464f9982-87db-4cc0-991d-83eb0148e61b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "464f9982-87db-4cc0-991d-83eb0148e61b" (UID: "464f9982-87db-4cc0-991d-83eb0148e61b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.156468 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/464f9982-87db-4cc0-991d-83eb0148e61b-kube-api-access-vk5h4" (OuterVolumeSpecName: "kube-api-access-vk5h4") pod "464f9982-87db-4cc0-991d-83eb0148e61b" (UID: "464f9982-87db-4cc0-991d-83eb0148e61b"). InnerVolumeSpecName "kube-api-access-vk5h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.156663 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-scripts" (OuterVolumeSpecName: "scripts") pod "464f9982-87db-4cc0-991d-83eb0148e61b" (UID: "464f9982-87db-4cc0-991d-83eb0148e61b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.156947 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "464f9982-87db-4cc0-991d-83eb0148e61b" (UID: "464f9982-87db-4cc0-991d-83eb0148e61b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.195655 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "464f9982-87db-4cc0-991d-83eb0148e61b" (UID: "464f9982-87db-4cc0-991d-83eb0148e61b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.212887 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data" (OuterVolumeSpecName: "config-data") pod "464f9982-87db-4cc0-991d-83eb0148e61b" (UID: "464f9982-87db-4cc0-991d-83eb0148e61b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.254307 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/464f9982-87db-4cc0-991d-83eb0148e61b-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.254341 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/464f9982-87db-4cc0-991d-83eb0148e61b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.254354 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk5h4\" (UniqueName: \"kubernetes.io/projected/464f9982-87db-4cc0-991d-83eb0148e61b-kube-api-access-vk5h4\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.254364 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.254372 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.254381 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.254390 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/464f9982-87db-4cc0-991d-83eb0148e61b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.423198 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"639c8ab0-c11a-4cf1-b944-9caa6b91604e","Type":"ContainerStarted","Data":"41538cf8b05d4931885d882441c42fce8934d4f6d669556929d964ce7c56df40"} Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.426267 4733 generic.go:334] "Generic (PLEG): container finished" podID="464f9982-87db-4cc0-991d-83eb0148e61b" containerID="5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1" exitCode=0 Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.426388 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.426491 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"464f9982-87db-4cc0-991d-83eb0148e61b","Type":"ContainerDied","Data":"5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1"} Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.426594 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"464f9982-87db-4cc0-991d-83eb0148e61b","Type":"ContainerDied","Data":"ddfc28af4ca6389988ec13a0270d776ef798255d7a17ae466c982b723241808f"} Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.426697 4733 scope.go:117] "RemoveContainer" containerID="5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.449924 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.717893081 podStartE2EDuration="3.449907861s" podCreationTimestamp="2025-12-04 19:17:45 +0000 UTC" firstStartedPulling="2025-12-04 19:17:46.42191615 +0000 UTC m=+5928.377277196" lastFinishedPulling="2025-12-04 19:17:47.15393093 +0000 UTC m=+5929.109291976" observedRunningTime="2025-12-04 19:17:48.445656277 +0000 UTC m=+5930.401017333" watchObservedRunningTime="2025-12-04 19:17:48.449907861 +0000 UTC m=+5930.405268907" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.458660 4733 scope.go:117] "RemoveContainer" containerID="ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.469709 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.477282 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.509118 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.513947 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:48 crc kubenswrapper[4733]: E1204 19:17:48.514543 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" containerName="cinder-api" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.514577 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" containerName="cinder-api" Dec 04 19:17:48 crc kubenswrapper[4733]: E1204 19:17:48.514618 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" containerName="cinder-api-log" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.514633 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" containerName="cinder-api-log" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.515554 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" containerName="cinder-api" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.515604 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" containerName="cinder-api-log" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.517656 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.521304 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.523340 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.529344 4733 scope.go:117] "RemoveContainer" containerID="5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1" Dec 04 19:17:48 crc kubenswrapper[4733]: E1204 19:17:48.529675 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1\": container with ID starting with 5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1 not found: ID does not exist" containerID="5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.529711 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1"} err="failed to get container status \"5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1\": rpc error: code = NotFound desc = could not find container \"5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1\": container with ID starting with 5eaa2dd71b81b050da1c84e7dc0c4adcd6c65d8a800b0d8d27fa3288b1ed57f1 not found: ID does not exist" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.529733 4733 scope.go:117] "RemoveContainer" containerID="ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d" Dec 04 19:17:48 crc kubenswrapper[4733]: E1204 19:17:48.530177 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d\": container with ID starting with ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d not found: ID does not exist" containerID="ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.530289 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d"} err="failed to get container status \"ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d\": rpc error: code = NotFound desc = could not find container \"ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d\": container with ID starting with ca0caa9d81147c310fafcbf7bc73312f93b5d949e3f7f130644bab2782cab69d not found: ID does not exist" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.664807 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-config-data-custom\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.665221 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.665252 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-scripts\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.665464 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e3c816c8-8307-4a11-a4ba-33615272c2f0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.665557 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-config-data\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.665619 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3c816c8-8307-4a11-a4ba-33615272c2f0-logs\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.666658 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdxsn\" (UniqueName: \"kubernetes.io/projected/e3c816c8-8307-4a11-a4ba-33615272c2f0-kube-api-access-hdxsn\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771124 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771213 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-scripts\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771336 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e3c816c8-8307-4a11-a4ba-33615272c2f0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771397 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-config-data\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771437 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3c816c8-8307-4a11-a4ba-33615272c2f0-logs\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771469 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e3c816c8-8307-4a11-a4ba-33615272c2f0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771475 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdxsn\" (UniqueName: \"kubernetes.io/projected/e3c816c8-8307-4a11-a4ba-33615272c2f0-kube-api-access-hdxsn\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771867 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-config-data-custom\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.771913 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3c816c8-8307-4a11-a4ba-33615272c2f0-logs\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.776933 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-scripts\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.780308 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-config-data-custom\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.782825 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-config-data\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.790179 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c816c8-8307-4a11-a4ba-33615272c2f0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.790894 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdxsn\" (UniqueName: \"kubernetes.io/projected/e3c816c8-8307-4a11-a4ba-33615272c2f0-kube-api-access-hdxsn\") pod \"cinder-api-0\" (UID: \"e3c816c8-8307-4a11-a4ba-33615272c2f0\") " pod="openstack/cinder-api-0" Dec 04 19:17:48 crc kubenswrapper[4733]: I1204 19:17:48.850309 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 19:17:49 crc kubenswrapper[4733]: I1204 19:17:49.302205 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 19:17:49 crc kubenswrapper[4733]: I1204 19:17:49.437836 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3c816c8-8307-4a11-a4ba-33615272c2f0","Type":"ContainerStarted","Data":"4f2f706277fbe41ec3b2a07e2b74c462f610de921aca05e67652cf8738a44268"} Dec 04 19:17:50 crc kubenswrapper[4733]: I1204 19:17:50.238496 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:50 crc kubenswrapper[4733]: I1204 19:17:50.363746 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="464f9982-87db-4cc0-991d-83eb0148e61b" path="/var/lib/kubelet/pods/464f9982-87db-4cc0-991d-83eb0148e61b/volumes" Dec 04 19:17:50 crc kubenswrapper[4733]: I1204 19:17:50.452729 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3c816c8-8307-4a11-a4ba-33615272c2f0","Type":"ContainerStarted","Data":"231c0a604855184957948d7392ba857fab5e7a11b664fab2d4012bb91fd41f31"} Dec 04 19:17:50 crc kubenswrapper[4733]: I1204 19:17:50.840125 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 04 19:17:51 crc kubenswrapper[4733]: I1204 19:17:51.463320 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e3c816c8-8307-4a11-a4ba-33615272c2f0","Type":"ContainerStarted","Data":"324bc2a2a3fb07b94932fbe70f05258ad07d33adbf49fbf874a7de1b866c0e64"} Dec 04 19:17:51 crc kubenswrapper[4733]: I1204 19:17:51.463496 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 19:17:51 crc kubenswrapper[4733]: I1204 19:17:51.490998 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.490977523 podStartE2EDuration="3.490977523s" podCreationTimestamp="2025-12-04 19:17:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:51.480955971 +0000 UTC m=+5933.436317047" watchObservedRunningTime="2025-12-04 19:17:51.490977523 +0000 UTC m=+5933.446338569" Dec 04 19:17:53 crc kubenswrapper[4733]: I1204 19:17:53.701684 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 19:17:53 crc kubenswrapper[4733]: I1204 19:17:53.773897 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:54 crc kubenswrapper[4733]: I1204 19:17:54.335610 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:17:54 crc kubenswrapper[4733]: E1204 19:17:54.335937 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:17:54 crc kubenswrapper[4733]: I1204 19:17:54.509283 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerName="cinder-scheduler" containerID="cri-o://28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87" gracePeriod=30 Dec 04 19:17:54 crc kubenswrapper[4733]: I1204 19:17:54.509428 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerName="probe" containerID="cri-o://44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23" gracePeriod=30 Dec 04 19:17:55 crc kubenswrapper[4733]: I1204 19:17:55.450148 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 04 19:17:55 crc kubenswrapper[4733]: I1204 19:17:55.530610 4733 generic.go:334] "Generic (PLEG): container finished" podID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerID="44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23" exitCode=0 Dec 04 19:17:55 crc kubenswrapper[4733]: I1204 19:17:55.530667 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60698b9c-8516-4023-a27c-e6bcaad6ee7a","Type":"ContainerDied","Data":"44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23"} Dec 04 19:17:55 crc kubenswrapper[4733]: I1204 19:17:55.991481 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.063275 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.115651 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-scripts\") pod \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.115740 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60698b9c-8516-4023-a27c-e6bcaad6ee7a-etc-machine-id\") pod \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.115826 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-combined-ca-bundle\") pod \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.115883 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2rjq\" (UniqueName: \"kubernetes.io/projected/60698b9c-8516-4023-a27c-e6bcaad6ee7a-kube-api-access-m2rjq\") pod \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.115945 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data\") pod \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.115974 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data-custom\") pod \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\" (UID: \"60698b9c-8516-4023-a27c-e6bcaad6ee7a\") " Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.116454 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60698b9c-8516-4023-a27c-e6bcaad6ee7a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "60698b9c-8516-4023-a27c-e6bcaad6ee7a" (UID: "60698b9c-8516-4023-a27c-e6bcaad6ee7a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.118587 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60698b9c-8516-4023-a27c-e6bcaad6ee7a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.133673 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "60698b9c-8516-4023-a27c-e6bcaad6ee7a" (UID: "60698b9c-8516-4023-a27c-e6bcaad6ee7a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.133789 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60698b9c-8516-4023-a27c-e6bcaad6ee7a-kube-api-access-m2rjq" (OuterVolumeSpecName: "kube-api-access-m2rjq") pod "60698b9c-8516-4023-a27c-e6bcaad6ee7a" (UID: "60698b9c-8516-4023-a27c-e6bcaad6ee7a"). InnerVolumeSpecName "kube-api-access-m2rjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.136771 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-scripts" (OuterVolumeSpecName: "scripts") pod "60698b9c-8516-4023-a27c-e6bcaad6ee7a" (UID: "60698b9c-8516-4023-a27c-e6bcaad6ee7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.170885 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60698b9c-8516-4023-a27c-e6bcaad6ee7a" (UID: "60698b9c-8516-4023-a27c-e6bcaad6ee7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.220647 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data" (OuterVolumeSpecName: "config-data") pod "60698b9c-8516-4023-a27c-e6bcaad6ee7a" (UID: "60698b9c-8516-4023-a27c-e6bcaad6ee7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.220851 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.220883 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.220900 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2rjq\" (UniqueName: \"kubernetes.io/projected/60698b9c-8516-4023-a27c-e6bcaad6ee7a-kube-api-access-m2rjq\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.220910 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.220918 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60698b9c-8516-4023-a27c-e6bcaad6ee7a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.542672 4733 generic.go:334] "Generic (PLEG): container finished" podID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerID="28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87" exitCode=0 Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.542714 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60698b9c-8516-4023-a27c-e6bcaad6ee7a","Type":"ContainerDied","Data":"28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87"} Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.542743 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60698b9c-8516-4023-a27c-e6bcaad6ee7a","Type":"ContainerDied","Data":"ec5434c1c3643a2c7cda864aa8ee6be7e9d9d3d7d80ca2b1b5bca0cf12e0c2a6"} Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.542767 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.542774 4733 scope.go:117] "RemoveContainer" containerID="44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.571389 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.574924 4733 scope.go:117] "RemoveContainer" containerID="28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.586333 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.596666 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:56 crc kubenswrapper[4733]: E1204 19:17:56.599140 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerName="cinder-scheduler" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.599346 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerName="cinder-scheduler" Dec 04 19:17:56 crc kubenswrapper[4733]: E1204 19:17:56.599433 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerName="probe" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.600145 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerName="probe" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.600477 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerName="cinder-scheduler" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.600611 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" containerName="probe" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.602039 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.604541 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.611423 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.624745 4733 scope.go:117] "RemoveContainer" containerID="44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23" Dec 04 19:17:56 crc kubenswrapper[4733]: E1204 19:17:56.629524 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23\": container with ID starting with 44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23 not found: ID does not exist" containerID="44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.629780 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23"} err="failed to get container status \"44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23\": rpc error: code = NotFound desc = could not find container \"44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23\": container with ID starting with 44b3bb4724802dbbac8806e17f6ea593a1e12018d8aab85884b2f4e147c35a23 not found: ID does not exist" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.629954 4733 scope.go:117] "RemoveContainer" containerID="28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.630039 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.630173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdgp8\" (UniqueName: \"kubernetes.io/projected/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-kube-api-access-wdgp8\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.630414 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-scripts\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.630506 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.630594 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.630726 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-config-data\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: E1204 19:17:56.631063 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87\": container with ID starting with 28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87 not found: ID does not exist" containerID="28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.631118 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87"} err="failed to get container status \"28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87\": rpc error: code = NotFound desc = could not find container \"28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87\": container with ID starting with 28f5dbd2f159c6a39a4f5b6dfcf4f4086a281fd8aeac5c3d6e611e3b5f899f87 not found: ID does not exist" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.732317 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdgp8\" (UniqueName: \"kubernetes.io/projected/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-kube-api-access-wdgp8\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.732418 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-scripts\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.732457 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.732492 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.732520 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-config-data\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.732557 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.732775 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.736870 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.737400 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.738163 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-config-data\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.741248 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-scripts\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.752960 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdgp8\" (UniqueName: \"kubernetes.io/projected/f32fb6b2-626f-4319-92cb-3a129ec9ec7d-kube-api-access-wdgp8\") pod \"cinder-scheduler-0\" (UID: \"f32fb6b2-626f-4319-92cb-3a129ec9ec7d\") " pod="openstack/cinder-scheduler-0" Dec 04 19:17:56 crc kubenswrapper[4733]: I1204 19:17:56.925776 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 19:17:57 crc kubenswrapper[4733]: I1204 19:17:57.396079 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 19:17:57 crc kubenswrapper[4733]: W1204 19:17:57.397284 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf32fb6b2_626f_4319_92cb_3a129ec9ec7d.slice/crio-6ecbc88ab2d10eee4c251b71aac614c5e0c925beaf60e7a974bc2a7e1a8b538f WatchSource:0}: Error finding container 6ecbc88ab2d10eee4c251b71aac614c5e0c925beaf60e7a974bc2a7e1a8b538f: Status 404 returned error can't find the container with id 6ecbc88ab2d10eee4c251b71aac614c5e0c925beaf60e7a974bc2a7e1a8b538f Dec 04 19:17:57 crc kubenswrapper[4733]: I1204 19:17:57.556728 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f32fb6b2-626f-4319-92cb-3a129ec9ec7d","Type":"ContainerStarted","Data":"6ecbc88ab2d10eee4c251b71aac614c5e0c925beaf60e7a974bc2a7e1a8b538f"} Dec 04 19:17:58 crc kubenswrapper[4733]: I1204 19:17:58.367970 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60698b9c-8516-4023-a27c-e6bcaad6ee7a" path="/var/lib/kubelet/pods/60698b9c-8516-4023-a27c-e6bcaad6ee7a/volumes" Dec 04 19:17:58 crc kubenswrapper[4733]: I1204 19:17:58.567234 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f32fb6b2-626f-4319-92cb-3a129ec9ec7d","Type":"ContainerStarted","Data":"fc9418f00613f984ba0f67577d32dde5163c152d5891ec00bdfca8af43228f98"} Dec 04 19:17:58 crc kubenswrapper[4733]: I1204 19:17:58.567456 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f32fb6b2-626f-4319-92cb-3a129ec9ec7d","Type":"ContainerStarted","Data":"fb6f89f9c15b38d6964d81555fa324eb9843ec216d69cc6cb70532552b13704a"} Dec 04 19:17:58 crc kubenswrapper[4733]: I1204 19:17:58.594610 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.594591303 podStartE2EDuration="2.594591303s" podCreationTimestamp="2025-12-04 19:17:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:17:58.589435893 +0000 UTC m=+5940.544796939" watchObservedRunningTime="2025-12-04 19:17:58.594591303 +0000 UTC m=+5940.549952349" Dec 04 19:18:00 crc kubenswrapper[4733]: I1204 19:18:00.704071 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 19:18:01 crc kubenswrapper[4733]: I1204 19:18:01.927289 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 19:18:07 crc kubenswrapper[4733]: I1204 19:18:07.137595 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 19:18:09 crc kubenswrapper[4733]: I1204 19:18:09.336455 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:18:09 crc kubenswrapper[4733]: E1204 19:18:09.337404 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:18:21 crc kubenswrapper[4733]: I1204 19:18:21.335753 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:18:21 crc kubenswrapper[4733]: E1204 19:18:21.337133 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:18:34 crc kubenswrapper[4733]: I1204 19:18:34.335513 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:18:34 crc kubenswrapper[4733]: E1204 19:18:34.336334 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:18:47 crc kubenswrapper[4733]: I1204 19:18:47.335316 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:18:47 crc kubenswrapper[4733]: E1204 19:18:47.336063 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:19:01 crc kubenswrapper[4733]: I1204 19:19:01.336694 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:19:01 crc kubenswrapper[4733]: E1204 19:19:01.337698 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:19:12 crc kubenswrapper[4733]: I1204 19:19:12.337008 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:19:12 crc kubenswrapper[4733]: E1204 19:19:12.338191 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:19:22 crc kubenswrapper[4733]: I1204 19:19:22.577685 4733 scope.go:117] "RemoveContainer" containerID="5044dc6f3f176ab12bcfd24b038c9b3e30c41bdfcca17ba0d9543b9ad36c61c1" Dec 04 19:19:22 crc kubenswrapper[4733]: I1204 19:19:22.620157 4733 scope.go:117] "RemoveContainer" containerID="0e8c013c0cd2969c07d97d46b59ea682500887d2a100cda7963f94cc793407ba" Dec 04 19:19:24 crc kubenswrapper[4733]: I1204 19:19:24.336159 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:19:24 crc kubenswrapper[4733]: E1204 19:19:24.336765 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.625437 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jdp84"] Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.627658 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.639944 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jdp84"] Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.717078 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-catalog-content\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.717137 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j585d\" (UniqueName: \"kubernetes.io/projected/6c99fc01-6ea0-42bf-8cda-ddc45b418545-kube-api-access-j585d\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.717332 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-utilities\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.818821 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j585d\" (UniqueName: \"kubernetes.io/projected/6c99fc01-6ea0-42bf-8cda-ddc45b418545-kube-api-access-j585d\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.819016 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-utilities\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.819064 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-catalog-content\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.819588 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-catalog-content\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.819639 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-utilities\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.852332 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j585d\" (UniqueName: \"kubernetes.io/projected/6c99fc01-6ea0-42bf-8cda-ddc45b418545-kube-api-access-j585d\") pod \"certified-operators-jdp84\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:34 crc kubenswrapper[4733]: I1204 19:19:34.967858 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:35 crc kubenswrapper[4733]: I1204 19:19:35.291958 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jdp84"] Dec 04 19:19:35 crc kubenswrapper[4733]: I1204 19:19:35.571908 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdp84" event={"ID":"6c99fc01-6ea0-42bf-8cda-ddc45b418545","Type":"ContainerStarted","Data":"427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512"} Dec 04 19:19:35 crc kubenswrapper[4733]: I1204 19:19:35.572211 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdp84" event={"ID":"6c99fc01-6ea0-42bf-8cda-ddc45b418545","Type":"ContainerStarted","Data":"29acc720b638c8f30090ebd19cba86f4afed46164e7f172c11e7b0184d920dd7"} Dec 04 19:19:36 crc kubenswrapper[4733]: I1204 19:19:36.580994 4733 generic.go:334] "Generic (PLEG): container finished" podID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerID="427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512" exitCode=0 Dec 04 19:19:36 crc kubenswrapper[4733]: I1204 19:19:36.581091 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdp84" event={"ID":"6c99fc01-6ea0-42bf-8cda-ddc45b418545","Type":"ContainerDied","Data":"427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512"} Dec 04 19:19:36 crc kubenswrapper[4733]: I1204 19:19:36.582976 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:19:37 crc kubenswrapper[4733]: I1204 19:19:37.597703 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdp84" event={"ID":"6c99fc01-6ea0-42bf-8cda-ddc45b418545","Type":"ContainerStarted","Data":"075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33"} Dec 04 19:19:38 crc kubenswrapper[4733]: I1204 19:19:38.610401 4733 generic.go:334] "Generic (PLEG): container finished" podID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerID="075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33" exitCode=0 Dec 04 19:19:38 crc kubenswrapper[4733]: I1204 19:19:38.610454 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdp84" event={"ID":"6c99fc01-6ea0-42bf-8cda-ddc45b418545","Type":"ContainerDied","Data":"075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33"} Dec 04 19:19:39 crc kubenswrapper[4733]: I1204 19:19:39.335400 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:19:39 crc kubenswrapper[4733]: E1204 19:19:39.335642 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:19:39 crc kubenswrapper[4733]: I1204 19:19:39.621290 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdp84" event={"ID":"6c99fc01-6ea0-42bf-8cda-ddc45b418545","Type":"ContainerStarted","Data":"4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3"} Dec 04 19:19:39 crc kubenswrapper[4733]: I1204 19:19:39.650023 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jdp84" podStartSLOduration=3.110873712 podStartE2EDuration="5.650006067s" podCreationTimestamp="2025-12-04 19:19:34 +0000 UTC" firstStartedPulling="2025-12-04 19:19:36.582644275 +0000 UTC m=+6038.538005321" lastFinishedPulling="2025-12-04 19:19:39.1217766 +0000 UTC m=+6041.077137676" observedRunningTime="2025-12-04 19:19:39.642143434 +0000 UTC m=+6041.597504480" watchObservedRunningTime="2025-12-04 19:19:39.650006067 +0000 UTC m=+6041.605367113" Dec 04 19:19:44 crc kubenswrapper[4733]: I1204 19:19:44.968759 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:44 crc kubenswrapper[4733]: I1204 19:19:44.969420 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:45 crc kubenswrapper[4733]: I1204 19:19:45.048861 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:45 crc kubenswrapper[4733]: I1204 19:19:45.720384 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:45 crc kubenswrapper[4733]: I1204 19:19:45.788040 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jdp84"] Dec 04 19:19:46 crc kubenswrapper[4733]: I1204 19:19:46.952491 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jnvbm"] Dec 04 19:19:46 crc kubenswrapper[4733]: I1204 19:19:46.956706 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:46 crc kubenswrapper[4733]: I1204 19:19:46.962428 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-rwfdr" Dec 04 19:19:46 crc kubenswrapper[4733]: I1204 19:19:46.963233 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 04 19:19:46 crc kubenswrapper[4733]: I1204 19:19:46.973736 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-h6kk4"] Dec 04 19:19:46 crc kubenswrapper[4733]: I1204 19:19:46.977180 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:46 crc kubenswrapper[4733]: I1204 19:19:46.991142 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jnvbm"] Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.006445 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-h6kk4"] Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-log\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100101 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25vq8\" (UniqueName: \"kubernetes.io/projected/0d759993-4494-4925-a9cf-97897f8c1383-kube-api-access-25vq8\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100139 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-etc-ovs\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100160 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-log-ovn\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100178 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d759993-4494-4925-a9cf-97897f8c1383-scripts\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100199 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-run\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100234 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ee5eee1-8eae-4de7-a2df-9ebbea833016-scripts\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100267 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-run\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100306 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-lib\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100330 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5mfz\" (UniqueName: \"kubernetes.io/projected/9ee5eee1-8eae-4de7-a2df-9ebbea833016-kube-api-access-x5mfz\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.100356 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-run-ovn\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202501 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-lib\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202577 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5mfz\" (UniqueName: \"kubernetes.io/projected/9ee5eee1-8eae-4de7-a2df-9ebbea833016-kube-api-access-x5mfz\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202620 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-run-ovn\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202669 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-log\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202706 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25vq8\" (UniqueName: \"kubernetes.io/projected/0d759993-4494-4925-a9cf-97897f8c1383-kube-api-access-25vq8\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202743 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-etc-ovs\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-log-ovn\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202785 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d759993-4494-4925-a9cf-97897f8c1383-scripts\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202823 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-run\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202861 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ee5eee1-8eae-4de7-a2df-9ebbea833016-scripts\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.202895 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-run\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.203030 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-run-ovn\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.203049 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-lib\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.203109 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-etc-ovs\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.203088 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-run\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.203136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0d759993-4494-4925-a9cf-97897f8c1383-var-log\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.203132 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-log-ovn\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.203220 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ee5eee1-8eae-4de7-a2df-9ebbea833016-var-run\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.205303 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d759993-4494-4925-a9cf-97897f8c1383-scripts\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.205660 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ee5eee1-8eae-4de7-a2df-9ebbea833016-scripts\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.223224 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5mfz\" (UniqueName: \"kubernetes.io/projected/9ee5eee1-8eae-4de7-a2df-9ebbea833016-kube-api-access-x5mfz\") pod \"ovn-controller-jnvbm\" (UID: \"9ee5eee1-8eae-4de7-a2df-9ebbea833016\") " pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.225361 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25vq8\" (UniqueName: \"kubernetes.io/projected/0d759993-4494-4925-a9cf-97897f8c1383-kube-api-access-25vq8\") pod \"ovn-controller-ovs-h6kk4\" (UID: \"0d759993-4494-4925-a9cf-97897f8c1383\") " pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.283202 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.299877 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.690904 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jdp84" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerName="registry-server" containerID="cri-o://4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3" gracePeriod=2 Dec 04 19:19:47 crc kubenswrapper[4733]: I1204 19:19:47.776507 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jnvbm"] Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.305370 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-h6kk4"] Dec 04 19:19:48 crc kubenswrapper[4733]: W1204 19:19:48.311958 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d759993_4494_4925_a9cf_97897f8c1383.slice/crio-df3994ecdb92c1def92a585b8ab1f677333599aa4a0a4d5d9851b188eae9de0e WatchSource:0}: Error finding container df3994ecdb92c1def92a585b8ab1f677333599aa4a0a4d5d9851b188eae9de0e: Status 404 returned error can't find the container with id df3994ecdb92c1def92a585b8ab1f677333599aa4a0a4d5d9851b188eae9de0e Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.415431 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.528910 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-catalog-content\") pod \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.529294 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-utilities\") pod \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.529363 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j585d\" (UniqueName: \"kubernetes.io/projected/6c99fc01-6ea0-42bf-8cda-ddc45b418545-kube-api-access-j585d\") pod \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\" (UID: \"6c99fc01-6ea0-42bf-8cda-ddc45b418545\") " Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.530752 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-utilities" (OuterVolumeSpecName: "utilities") pod "6c99fc01-6ea0-42bf-8cda-ddc45b418545" (UID: "6c99fc01-6ea0-42bf-8cda-ddc45b418545"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.540522 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c99fc01-6ea0-42bf-8cda-ddc45b418545-kube-api-access-j585d" (OuterVolumeSpecName: "kube-api-access-j585d") pod "6c99fc01-6ea0-42bf-8cda-ddc45b418545" (UID: "6c99fc01-6ea0-42bf-8cda-ddc45b418545"). InnerVolumeSpecName "kube-api-access-j585d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.585159 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c99fc01-6ea0-42bf-8cda-ddc45b418545" (UID: "6c99fc01-6ea0-42bf-8cda-ddc45b418545"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.631079 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.631108 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c99fc01-6ea0-42bf-8cda-ddc45b418545-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.631118 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j585d\" (UniqueName: \"kubernetes.io/projected/6c99fc01-6ea0-42bf-8cda-ddc45b418545-kube-api-access-j585d\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.703329 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jnvbm" event={"ID":"9ee5eee1-8eae-4de7-a2df-9ebbea833016","Type":"ContainerStarted","Data":"dec5071c5ccc0490267089b64f3746096de3ef9fa035aa366cf45e067ca129ed"} Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.703407 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jnvbm" event={"ID":"9ee5eee1-8eae-4de7-a2df-9ebbea833016","Type":"ContainerStarted","Data":"15b776d87b0325b05b5bbb63fd953c91181a1b7aabf4b7788284415704e0b9cf"} Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.703862 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jnvbm" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.705233 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-h6kk4" event={"ID":"0d759993-4494-4925-a9cf-97897f8c1383","Type":"ContainerStarted","Data":"0dc0fbcf0f70f338b02364528e2f22e83f1aea100df0edb22f96156729d63ec9"} Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.705260 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-h6kk4" event={"ID":"0d759993-4494-4925-a9cf-97897f8c1383","Type":"ContainerStarted","Data":"df3994ecdb92c1def92a585b8ab1f677333599aa4a0a4d5d9851b188eae9de0e"} Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.707983 4733 generic.go:334] "Generic (PLEG): container finished" podID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerID="4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3" exitCode=0 Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.708014 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdp84" event={"ID":"6c99fc01-6ea0-42bf-8cda-ddc45b418545","Type":"ContainerDied","Data":"4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3"} Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.708031 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdp84" event={"ID":"6c99fc01-6ea0-42bf-8cda-ddc45b418545","Type":"ContainerDied","Data":"29acc720b638c8f30090ebd19cba86f4afed46164e7f172c11e7b0184d920dd7"} Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.708051 4733 scope.go:117] "RemoveContainer" containerID="4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.708167 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jdp84" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.736816 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jnvbm" podStartSLOduration=2.736777985 podStartE2EDuration="2.736777985s" podCreationTimestamp="2025-12-04 19:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:19:48.728852691 +0000 UTC m=+6050.684213747" watchObservedRunningTime="2025-12-04 19:19:48.736777985 +0000 UTC m=+6050.692139031" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.743737 4733 scope.go:117] "RemoveContainer" containerID="075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.824681 4733 scope.go:117] "RemoveContainer" containerID="427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.832193 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jdp84"] Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.841479 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jdp84"] Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.873049 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-qfl9v"] Dec 04 19:19:48 crc kubenswrapper[4733]: E1204 19:19:48.873605 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerName="extract-content" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.873621 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerName="extract-content" Dec 04 19:19:48 crc kubenswrapper[4733]: E1204 19:19:48.873649 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerName="extract-utilities" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.873658 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerName="extract-utilities" Dec 04 19:19:48 crc kubenswrapper[4733]: E1204 19:19:48.873685 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerName="registry-server" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.873694 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerName="registry-server" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.874033 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" containerName="registry-server" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.874879 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.876210 4733 scope.go:117] "RemoveContainer" containerID="4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3" Dec 04 19:19:48 crc kubenswrapper[4733]: E1204 19:19:48.876966 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3\": container with ID starting with 4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3 not found: ID does not exist" containerID="4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.876995 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3"} err="failed to get container status \"4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3\": rpc error: code = NotFound desc = could not find container \"4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3\": container with ID starting with 4a95cddf122e5ec45569a95dd10d1087a00f86568df0d5dc512ff694f486dab3 not found: ID does not exist" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.877016 4733 scope.go:117] "RemoveContainer" containerID="075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33" Dec 04 19:19:48 crc kubenswrapper[4733]: E1204 19:19:48.877210 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33\": container with ID starting with 075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33 not found: ID does not exist" containerID="075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.877232 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33"} err="failed to get container status \"075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33\": rpc error: code = NotFound desc = could not find container \"075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33\": container with ID starting with 075286736a9908b88a7222b4970a7c84faa90d194641208628cf33d512968d33 not found: ID does not exist" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.877244 4733 scope.go:117] "RemoveContainer" containerID="427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512" Dec 04 19:19:48 crc kubenswrapper[4733]: E1204 19:19:48.877429 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512\": container with ID starting with 427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512 not found: ID does not exist" containerID="427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.877450 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512"} err="failed to get container status \"427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512\": rpc error: code = NotFound desc = could not find container \"427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512\": container with ID starting with 427f9ea9baba508ea75a800f5f9bef8e0094051363dcc602d705cd2fc696d512 not found: ID does not exist" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.883278 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-qfl9v"] Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.938204 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzl4t\" (UniqueName: \"kubernetes.io/projected/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-kube-api-access-kzl4t\") pod \"octavia-db-create-qfl9v\" (UID: \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\") " pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:48 crc kubenswrapper[4733]: I1204 19:19:48.938539 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-operator-scripts\") pod \"octavia-db-create-qfl9v\" (UID: \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\") " pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.040485 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzl4t\" (UniqueName: \"kubernetes.io/projected/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-kube-api-access-kzl4t\") pod \"octavia-db-create-qfl9v\" (UID: \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\") " pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.040538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-operator-scripts\") pod \"octavia-db-create-qfl9v\" (UID: \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\") " pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.041220 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-operator-scripts\") pod \"octavia-db-create-qfl9v\" (UID: \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\") " pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.087218 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzl4t\" (UniqueName: \"kubernetes.io/projected/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-kube-api-access-kzl4t\") pod \"octavia-db-create-qfl9v\" (UID: \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\") " pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.199850 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.614357 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-wwr9r"] Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.615954 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.622858 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.634368 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-wwr9r"] Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.653140 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-ovn-rundir\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.653293 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9dbc\" (UniqueName: \"kubernetes.io/projected/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-kube-api-access-j9dbc\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.653335 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-ovs-rundir\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.653393 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-config\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.727505 4733 generic.go:334] "Generic (PLEG): container finished" podID="0d759993-4494-4925-a9cf-97897f8c1383" containerID="0dc0fbcf0f70f338b02364528e2f22e83f1aea100df0edb22f96156729d63ec9" exitCode=0 Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.727590 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-h6kk4" event={"ID":"0d759993-4494-4925-a9cf-97897f8c1383","Type":"ContainerDied","Data":"0dc0fbcf0f70f338b02364528e2f22e83f1aea100df0edb22f96156729d63ec9"} Dec 04 19:19:49 crc kubenswrapper[4733]: W1204 19:19:49.745890 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d8dacc3_457b_4cbc_b8c8_56ac4e333e17.slice/crio-b32543d2454b0d4266b92ec4e4a001622e2c6ae093459ceefe0661d897b31aa6 WatchSource:0}: Error finding container b32543d2454b0d4266b92ec4e4a001622e2c6ae093459ceefe0661d897b31aa6: Status 404 returned error can't find the container with id b32543d2454b0d4266b92ec4e4a001622e2c6ae093459ceefe0661d897b31aa6 Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.754709 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9dbc\" (UniqueName: \"kubernetes.io/projected/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-kube-api-access-j9dbc\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.754753 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-ovs-rundir\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.754807 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-config\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.754863 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-ovn-rundir\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.755166 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-ovn-rundir\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.755335 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-ovs-rundir\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.755816 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-config\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.763810 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-qfl9v"] Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.774659 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9dbc\" (UniqueName: \"kubernetes.io/projected/5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1-kube-api-access-j9dbc\") pod \"ovn-controller-metrics-wwr9r\" (UID: \"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1\") " pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:49 crc kubenswrapper[4733]: I1204 19:19:49.942174 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-wwr9r" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.347939 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c99fc01-6ea0-42bf-8cda-ddc45b418545" path="/var/lib/kubelet/pods/6c99fc01-6ea0-42bf-8cda-ddc45b418545/volumes" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.432452 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-wwr9r"] Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.739324 4733 generic.go:334] "Generic (PLEG): container finished" podID="1d8dacc3-457b-4cbc-b8c8-56ac4e333e17" containerID="9b1e2cbd8b82c3ecda12f0035b6050e8539ff39af8334049c19df65422266c2e" exitCode=0 Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.739492 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-qfl9v" event={"ID":"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17","Type":"ContainerDied","Data":"9b1e2cbd8b82c3ecda12f0035b6050e8539ff39af8334049c19df65422266c2e"} Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.739775 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-qfl9v" event={"ID":"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17","Type":"ContainerStarted","Data":"b32543d2454b0d4266b92ec4e4a001622e2c6ae093459ceefe0661d897b31aa6"} Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.741781 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-wwr9r" event={"ID":"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1","Type":"ContainerStarted","Data":"8c1a38e20e66b0aef128f0e181a622d886fe19a1bf99be7f5a696b704b996ec2"} Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.741829 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-wwr9r" event={"ID":"5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1","Type":"ContainerStarted","Data":"ebeea39a71b43e2275a74f3abb24d0333ff766f9315c19afaf5c85408d255cb6"} Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.744810 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-h6kk4" event={"ID":"0d759993-4494-4925-a9cf-97897f8c1383","Type":"ContainerStarted","Data":"b049939068692f4e095de85244a56a11ec94cc337b3d46805badfd3fffc2c732"} Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.744837 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-h6kk4" event={"ID":"0d759993-4494-4925-a9cf-97897f8c1383","Type":"ContainerStarted","Data":"f7b8ed17b685bf9824fc492df75613d64951dead27e4fc72b7f7d8fbfa54d16e"} Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.745439 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.745469 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.773251 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-wwr9r" podStartSLOduration=1.773232125 podStartE2EDuration="1.773232125s" podCreationTimestamp="2025-12-04 19:19:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:19:50.772223587 +0000 UTC m=+6052.727584633" watchObservedRunningTime="2025-12-04 19:19:50.773232125 +0000 UTC m=+6052.728593161" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.805225 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-h6kk4" podStartSLOduration=4.805204389 podStartE2EDuration="4.805204389s" podCreationTimestamp="2025-12-04 19:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:19:50.798039565 +0000 UTC m=+6052.753400631" watchObservedRunningTime="2025-12-04 19:19:50.805204389 +0000 UTC m=+6052.760565435" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.820706 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-c6b6-account-create-update-f94nx"] Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.821841 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.831694 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.841366 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-c6b6-account-create-update-f94nx"] Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.888538 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ldq9\" (UniqueName: \"kubernetes.io/projected/5023d9fb-76ff-4362-b564-c54737351d65-kube-api-access-5ldq9\") pod \"octavia-c6b6-account-create-update-f94nx\" (UID: \"5023d9fb-76ff-4362-b564-c54737351d65\") " pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.888692 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5023d9fb-76ff-4362-b564-c54737351d65-operator-scripts\") pod \"octavia-c6b6-account-create-update-f94nx\" (UID: \"5023d9fb-76ff-4362-b564-c54737351d65\") " pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.990979 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5023d9fb-76ff-4362-b564-c54737351d65-operator-scripts\") pod \"octavia-c6b6-account-create-update-f94nx\" (UID: \"5023d9fb-76ff-4362-b564-c54737351d65\") " pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.991077 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ldq9\" (UniqueName: \"kubernetes.io/projected/5023d9fb-76ff-4362-b564-c54737351d65-kube-api-access-5ldq9\") pod \"octavia-c6b6-account-create-update-f94nx\" (UID: \"5023d9fb-76ff-4362-b564-c54737351d65\") " pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:50 crc kubenswrapper[4733]: I1204 19:19:50.991949 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5023d9fb-76ff-4362-b564-c54737351d65-operator-scripts\") pod \"octavia-c6b6-account-create-update-f94nx\" (UID: \"5023d9fb-76ff-4362-b564-c54737351d65\") " pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:51 crc kubenswrapper[4733]: I1204 19:19:51.022753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ldq9\" (UniqueName: \"kubernetes.io/projected/5023d9fb-76ff-4362-b564-c54737351d65-kube-api-access-5ldq9\") pod \"octavia-c6b6-account-create-update-f94nx\" (UID: \"5023d9fb-76ff-4362-b564-c54737351d65\") " pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:51 crc kubenswrapper[4733]: I1204 19:19:51.151883 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:51 crc kubenswrapper[4733]: W1204 19:19:51.613747 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5023d9fb_76ff_4362_b564_c54737351d65.slice/crio-cc54dcba128e309346b68cb4d1b7ef90bba237930a96a46ad504dedd5ada9172 WatchSource:0}: Error finding container cc54dcba128e309346b68cb4d1b7ef90bba237930a96a46ad504dedd5ada9172: Status 404 returned error can't find the container with id cc54dcba128e309346b68cb4d1b7ef90bba237930a96a46ad504dedd5ada9172 Dec 04 19:19:51 crc kubenswrapper[4733]: I1204 19:19:51.618970 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-c6b6-account-create-update-f94nx"] Dec 04 19:19:51 crc kubenswrapper[4733]: I1204 19:19:51.757048 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c6b6-account-create-update-f94nx" event={"ID":"5023d9fb-76ff-4362-b564-c54737351d65","Type":"ContainerStarted","Data":"cc54dcba128e309346b68cb4d1b7ef90bba237930a96a46ad504dedd5ada9172"} Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.114941 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.215616 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzl4t\" (UniqueName: \"kubernetes.io/projected/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-kube-api-access-kzl4t\") pod \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\" (UID: \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\") " Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.215668 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-operator-scripts\") pod \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\" (UID: \"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17\") " Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.217128 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d8dacc3-457b-4cbc-b8c8-56ac4e333e17" (UID: "1d8dacc3-457b-4cbc-b8c8-56ac4e333e17"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.221083 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-kube-api-access-kzl4t" (OuterVolumeSpecName: "kube-api-access-kzl4t") pod "1d8dacc3-457b-4cbc-b8c8-56ac4e333e17" (UID: "1d8dacc3-457b-4cbc-b8c8-56ac4e333e17"). InnerVolumeSpecName "kube-api-access-kzl4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.318277 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzl4t\" (UniqueName: \"kubernetes.io/projected/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-kube-api-access-kzl4t\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.318312 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.788646 4733 generic.go:334] "Generic (PLEG): container finished" podID="5023d9fb-76ff-4362-b564-c54737351d65" containerID="80a849d764c8dbfb9752cfdef552031d425786c4297340190f3d973b5de8efe2" exitCode=0 Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.788904 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c6b6-account-create-update-f94nx" event={"ID":"5023d9fb-76ff-4362-b564-c54737351d65","Type":"ContainerDied","Data":"80a849d764c8dbfb9752cfdef552031d425786c4297340190f3d973b5de8efe2"} Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.797668 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-qfl9v" Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.797994 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-qfl9v" event={"ID":"1d8dacc3-457b-4cbc-b8c8-56ac4e333e17","Type":"ContainerDied","Data":"b32543d2454b0d4266b92ec4e4a001622e2c6ae093459ceefe0661d897b31aa6"} Dec 04 19:19:52 crc kubenswrapper[4733]: I1204 19:19:52.798045 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b32543d2454b0d4266b92ec4e4a001622e2c6ae093459ceefe0661d897b31aa6" Dec 04 19:19:53 crc kubenswrapper[4733]: I1204 19:19:53.335577 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:19:53 crc kubenswrapper[4733]: E1204 19:19:53.335842 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.080749 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f755-account-create-update-sgvzv"] Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.092402 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f755-account-create-update-sgvzv"] Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.102571 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bsb7q"] Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.110983 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bsb7q"] Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.156822 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.255424 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ldq9\" (UniqueName: \"kubernetes.io/projected/5023d9fb-76ff-4362-b564-c54737351d65-kube-api-access-5ldq9\") pod \"5023d9fb-76ff-4362-b564-c54737351d65\" (UID: \"5023d9fb-76ff-4362-b564-c54737351d65\") " Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.255504 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5023d9fb-76ff-4362-b564-c54737351d65-operator-scripts\") pod \"5023d9fb-76ff-4362-b564-c54737351d65\" (UID: \"5023d9fb-76ff-4362-b564-c54737351d65\") " Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.256271 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5023d9fb-76ff-4362-b564-c54737351d65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5023d9fb-76ff-4362-b564-c54737351d65" (UID: "5023d9fb-76ff-4362-b564-c54737351d65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.260556 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5023d9fb-76ff-4362-b564-c54737351d65-kube-api-access-5ldq9" (OuterVolumeSpecName: "kube-api-access-5ldq9") pod "5023d9fb-76ff-4362-b564-c54737351d65" (UID: "5023d9fb-76ff-4362-b564-c54737351d65"). InnerVolumeSpecName "kube-api-access-5ldq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.348429 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f99dc23-04af-4f68-b2cc-7f9b5f29798d" path="/var/lib/kubelet/pods/4f99dc23-04af-4f68-b2cc-7f9b5f29798d/volumes" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.349296 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e23e0f7f-8466-4896-9aae-0fd66d607599" path="/var/lib/kubelet/pods/e23e0f7f-8466-4896-9aae-0fd66d607599/volumes" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.357587 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ldq9\" (UniqueName: \"kubernetes.io/projected/5023d9fb-76ff-4362-b564-c54737351d65-kube-api-access-5ldq9\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.357615 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5023d9fb-76ff-4362-b564-c54737351d65-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.817431 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c6b6-account-create-update-f94nx" event={"ID":"5023d9fb-76ff-4362-b564-c54737351d65","Type":"ContainerDied","Data":"cc54dcba128e309346b68cb4d1b7ef90bba237930a96a46ad504dedd5ada9172"} Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.817744 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc54dcba128e309346b68cb4d1b7ef90bba237930a96a46ad504dedd5ada9172" Dec 04 19:19:54 crc kubenswrapper[4733]: I1204 19:19:54.817490 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c6b6-account-create-update-f94nx" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.378650 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-6lknq"] Dec 04 19:19:56 crc kubenswrapper[4733]: E1204 19:19:56.380557 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5023d9fb-76ff-4362-b564-c54737351d65" containerName="mariadb-account-create-update" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.380692 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5023d9fb-76ff-4362-b564-c54737351d65" containerName="mariadb-account-create-update" Dec 04 19:19:56 crc kubenswrapper[4733]: E1204 19:19:56.381126 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8dacc3-457b-4cbc-b8c8-56ac4e333e17" containerName="mariadb-database-create" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.381247 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8dacc3-457b-4cbc-b8c8-56ac4e333e17" containerName="mariadb-database-create" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.381534 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5023d9fb-76ff-4362-b564-c54737351d65" containerName="mariadb-account-create-update" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.381612 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8dacc3-457b-4cbc-b8c8-56ac4e333e17" containerName="mariadb-database-create" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.382323 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.391431 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-6lknq"] Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.500031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-operator-scripts\") pod \"octavia-persistence-db-create-6lknq\" (UID: \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\") " pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.500467 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8s5s\" (UniqueName: \"kubernetes.io/projected/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-kube-api-access-k8s5s\") pod \"octavia-persistence-db-create-6lknq\" (UID: \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\") " pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.601767 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8s5s\" (UniqueName: \"kubernetes.io/projected/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-kube-api-access-k8s5s\") pod \"octavia-persistence-db-create-6lknq\" (UID: \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\") " pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.601859 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-operator-scripts\") pod \"octavia-persistence-db-create-6lknq\" (UID: \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\") " pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.602541 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-operator-scripts\") pod \"octavia-persistence-db-create-6lknq\" (UID: \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\") " pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.623718 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8s5s\" (UniqueName: \"kubernetes.io/projected/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-kube-api-access-k8s5s\") pod \"octavia-persistence-db-create-6lknq\" (UID: \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\") " pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:56 crc kubenswrapper[4733]: I1204 19:19:56.702236 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.297170 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-6lknq"] Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.478830 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-7a0a-account-create-update-45wx9"] Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.480278 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.482699 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.492459 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-7a0a-account-create-update-45wx9"] Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.624238 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km5bf\" (UniqueName: \"kubernetes.io/projected/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-kube-api-access-km5bf\") pod \"octavia-7a0a-account-create-update-45wx9\" (UID: \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\") " pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.624362 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-operator-scripts\") pod \"octavia-7a0a-account-create-update-45wx9\" (UID: \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\") " pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.726143 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km5bf\" (UniqueName: \"kubernetes.io/projected/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-kube-api-access-km5bf\") pod \"octavia-7a0a-account-create-update-45wx9\" (UID: \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\") " pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.726244 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-operator-scripts\") pod \"octavia-7a0a-account-create-update-45wx9\" (UID: \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\") " pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.726941 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-operator-scripts\") pod \"octavia-7a0a-account-create-update-45wx9\" (UID: \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\") " pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.751930 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km5bf\" (UniqueName: \"kubernetes.io/projected/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-kube-api-access-km5bf\") pod \"octavia-7a0a-account-create-update-45wx9\" (UID: \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\") " pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.802188 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.854972 4733 generic.go:334] "Generic (PLEG): container finished" podID="27cc81ff-912a-46e8-94ad-3a0dfd3a9085" containerID="6a500f5d0dce38480ed2ff82385fed20f4fabffc2e397f8601d3ec631f310d02" exitCode=0 Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.855194 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-6lknq" event={"ID":"27cc81ff-912a-46e8-94ad-3a0dfd3a9085","Type":"ContainerDied","Data":"6a500f5d0dce38480ed2ff82385fed20f4fabffc2e397f8601d3ec631f310d02"} Dec 04 19:19:57 crc kubenswrapper[4733]: I1204 19:19:57.855263 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-6lknq" event={"ID":"27cc81ff-912a-46e8-94ad-3a0dfd3a9085","Type":"ContainerStarted","Data":"8fbf117693819c113eff2354c75f6bd9f41d6cafc1696d16841746e8aaf359a4"} Dec 04 19:19:58 crc kubenswrapper[4733]: I1204 19:19:58.306215 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-7a0a-account-create-update-45wx9"] Dec 04 19:19:58 crc kubenswrapper[4733]: I1204 19:19:58.865447 4733 generic.go:334] "Generic (PLEG): container finished" podID="14a62c89-bc73-48b3-9bdf-775f5a3f9bdd" containerID="68f2489219caec47d0bf068e927c8f7687dd938ce8b6887f37c6f1e139abeb85" exitCode=0 Dec 04 19:19:58 crc kubenswrapper[4733]: I1204 19:19:58.865510 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-7a0a-account-create-update-45wx9" event={"ID":"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd","Type":"ContainerDied","Data":"68f2489219caec47d0bf068e927c8f7687dd938ce8b6887f37c6f1e139abeb85"} Dec 04 19:19:58 crc kubenswrapper[4733]: I1204 19:19:58.866110 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-7a0a-account-create-update-45wx9" event={"ID":"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd","Type":"ContainerStarted","Data":"6deabde354f19c7fc314ba99ad0746b1bc379131307cd82de55b317a05d30bf6"} Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.263939 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.385762 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8s5s\" (UniqueName: \"kubernetes.io/projected/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-kube-api-access-k8s5s\") pod \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\" (UID: \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\") " Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.385934 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-operator-scripts\") pod \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\" (UID: \"27cc81ff-912a-46e8-94ad-3a0dfd3a9085\") " Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.386560 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "27cc81ff-912a-46e8-94ad-3a0dfd3a9085" (UID: "27cc81ff-912a-46e8-94ad-3a0dfd3a9085"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.391590 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-kube-api-access-k8s5s" (OuterVolumeSpecName: "kube-api-access-k8s5s") pod "27cc81ff-912a-46e8-94ad-3a0dfd3a9085" (UID: "27cc81ff-912a-46e8-94ad-3a0dfd3a9085"). InnerVolumeSpecName "kube-api-access-k8s5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.488997 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.489088 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8s5s\" (UniqueName: \"kubernetes.io/projected/27cc81ff-912a-46e8-94ad-3a0dfd3a9085-kube-api-access-k8s5s\") on node \"crc\" DevicePath \"\"" Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.878079 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-6lknq" event={"ID":"27cc81ff-912a-46e8-94ad-3a0dfd3a9085","Type":"ContainerDied","Data":"8fbf117693819c113eff2354c75f6bd9f41d6cafc1696d16841746e8aaf359a4"} Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.878361 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fbf117693819c113eff2354c75f6bd9f41d6cafc1696d16841746e8aaf359a4" Dec 04 19:19:59 crc kubenswrapper[4733]: I1204 19:19:59.878125 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-6lknq" Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.306016 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.406122 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km5bf\" (UniqueName: \"kubernetes.io/projected/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-kube-api-access-km5bf\") pod \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\" (UID: \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\") " Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.407759 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-operator-scripts\") pod \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\" (UID: \"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd\") " Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.410591 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "14a62c89-bc73-48b3-9bdf-775f5a3f9bdd" (UID: "14a62c89-bc73-48b3-9bdf-775f5a3f9bdd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.410924 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-kube-api-access-km5bf" (OuterVolumeSpecName: "kube-api-access-km5bf") pod "14a62c89-bc73-48b3-9bdf-775f5a3f9bdd" (UID: "14a62c89-bc73-48b3-9bdf-775f5a3f9bdd"). InnerVolumeSpecName "kube-api-access-km5bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.510465 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km5bf\" (UniqueName: \"kubernetes.io/projected/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-kube-api-access-km5bf\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.510507 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.893878 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-7a0a-account-create-update-45wx9" event={"ID":"14a62c89-bc73-48b3-9bdf-775f5a3f9bdd","Type":"ContainerDied","Data":"6deabde354f19c7fc314ba99ad0746b1bc379131307cd82de55b317a05d30bf6"} Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.893938 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6deabde354f19c7fc314ba99ad0746b1bc379131307cd82de55b317a05d30bf6" Dec 04 19:20:00 crc kubenswrapper[4733]: I1204 19:20:00.893999 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-7a0a-account-create-update-45wx9" Dec 04 19:20:01 crc kubenswrapper[4733]: I1204 19:20:01.028660 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-sxc9r"] Dec 04 19:20:01 crc kubenswrapper[4733]: I1204 19:20:01.040785 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-sxc9r"] Dec 04 19:20:02 crc kubenswrapper[4733]: I1204 19:20:02.354544 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f77211d-e7c6-4a42-afd5-663e4a4170d2" path="/var/lib/kubelet/pods/4f77211d-e7c6-4a42-afd5-663e4a4170d2/volumes" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.469574 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-59457b6897-wl54k"] Dec 04 19:20:03 crc kubenswrapper[4733]: E1204 19:20:03.470375 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a62c89-bc73-48b3-9bdf-775f5a3f9bdd" containerName="mariadb-account-create-update" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.470391 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a62c89-bc73-48b3-9bdf-775f5a3f9bdd" containerName="mariadb-account-create-update" Dec 04 19:20:03 crc kubenswrapper[4733]: E1204 19:20:03.470441 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cc81ff-912a-46e8-94ad-3a0dfd3a9085" containerName="mariadb-database-create" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.470451 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cc81ff-912a-46e8-94ad-3a0dfd3a9085" containerName="mariadb-database-create" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.470660 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a62c89-bc73-48b3-9bdf-775f5a3f9bdd" containerName="mariadb-account-create-update" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.470688 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cc81ff-912a-46e8-94ad-3a0dfd3a9085" containerName="mariadb-database-create" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.475341 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.487336 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-59457b6897-wl54k"] Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.495903 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.495936 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.496059 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-szhdc" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.589894 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a777e880-541c-4aae-8249-d2027129051f-octavia-run\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.589979 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a777e880-541c-4aae-8249-d2027129051f-config-data-merged\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.590089 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-scripts\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.590121 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-combined-ca-bundle\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.590210 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-config-data\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.691896 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-config-data\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.691966 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a777e880-541c-4aae-8249-d2027129051f-octavia-run\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.692014 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a777e880-541c-4aae-8249-d2027129051f-config-data-merged\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.692092 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-scripts\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.692123 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-combined-ca-bundle\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.692527 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a777e880-541c-4aae-8249-d2027129051f-octavia-run\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.692819 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a777e880-541c-4aae-8249-d2027129051f-config-data-merged\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.697208 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-combined-ca-bundle\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.697433 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-config-data\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.704454 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a777e880-541c-4aae-8249-d2027129051f-scripts\") pod \"octavia-api-59457b6897-wl54k\" (UID: \"a777e880-541c-4aae-8249-d2027129051f\") " pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:03 crc kubenswrapper[4733]: I1204 19:20:03.827349 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:04 crc kubenswrapper[4733]: I1204 19:20:04.321591 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-59457b6897-wl54k"] Dec 04 19:20:04 crc kubenswrapper[4733]: W1204 19:20:04.324032 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda777e880_541c_4aae_8249_d2027129051f.slice/crio-0c215da48ffdd4c22b997a4097fcfb562f7922ea377162b04d811784e090a72e WatchSource:0}: Error finding container 0c215da48ffdd4c22b997a4097fcfb562f7922ea377162b04d811784e090a72e: Status 404 returned error can't find the container with id 0c215da48ffdd4c22b997a4097fcfb562f7922ea377162b04d811784e090a72e Dec 04 19:20:04 crc kubenswrapper[4733]: I1204 19:20:04.939490 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-59457b6897-wl54k" event={"ID":"a777e880-541c-4aae-8249-d2027129051f","Type":"ContainerStarted","Data":"0c215da48ffdd4c22b997a4097fcfb562f7922ea377162b04d811784e090a72e"} Dec 04 19:20:05 crc kubenswrapper[4733]: I1204 19:20:05.336442 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:20:05 crc kubenswrapper[4733]: E1204 19:20:05.336980 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:20:14 crc kubenswrapper[4733]: I1204 19:20:14.057360 4733 generic.go:334] "Generic (PLEG): container finished" podID="a777e880-541c-4aae-8249-d2027129051f" containerID="e9014930de26ba17e138022062bfd9ccf280c1c6ffa984c0e0eb8968acba5e90" exitCode=0 Dec 04 19:20:14 crc kubenswrapper[4733]: I1204 19:20:14.057899 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-59457b6897-wl54k" event={"ID":"a777e880-541c-4aae-8249-d2027129051f","Type":"ContainerDied","Data":"e9014930de26ba17e138022062bfd9ccf280c1c6ffa984c0e0eb8968acba5e90"} Dec 04 19:20:14 crc kubenswrapper[4733]: I1204 19:20:14.058214 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-m5tgq"] Dec 04 19:20:14 crc kubenswrapper[4733]: I1204 19:20:14.079750 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-m5tgq"] Dec 04 19:20:14 crc kubenswrapper[4733]: I1204 19:20:14.349229 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09d2f0ad-ae4c-4307-97d8-35ac282147bc" path="/var/lib/kubelet/pods/09d2f0ad-ae4c-4307-97d8-35ac282147bc/volumes" Dec 04 19:20:15 crc kubenswrapper[4733]: I1204 19:20:15.068476 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-59457b6897-wl54k" event={"ID":"a777e880-541c-4aae-8249-d2027129051f","Type":"ContainerStarted","Data":"a715a1a6cc80d25b6a0675a26272974e600eb3a81442296b6e8bb3d725dd7c98"} Dec 04 19:20:15 crc kubenswrapper[4733]: I1204 19:20:15.068724 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-59457b6897-wl54k" event={"ID":"a777e880-541c-4aae-8249-d2027129051f","Type":"ContainerStarted","Data":"1e9a6ef68e6f3c87ee47eb91c6163c66ad1449c79bc9551e5d5e9100d39d0a6b"} Dec 04 19:20:15 crc kubenswrapper[4733]: I1204 19:20:15.068742 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:15 crc kubenswrapper[4733]: I1204 19:20:15.089774 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-59457b6897-wl54k" podStartSLOduration=3.019657432 podStartE2EDuration="12.0897567s" podCreationTimestamp="2025-12-04 19:20:03 +0000 UTC" firstStartedPulling="2025-12-04 19:20:04.325772817 +0000 UTC m=+6066.281133853" lastFinishedPulling="2025-12-04 19:20:13.395872035 +0000 UTC m=+6075.351233121" observedRunningTime="2025-12-04 19:20:15.08643865 +0000 UTC m=+6077.041799696" watchObservedRunningTime="2025-12-04 19:20:15.0897567 +0000 UTC m=+6077.045117746" Dec 04 19:20:16 crc kubenswrapper[4733]: I1204 19:20:16.078209 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:17 crc kubenswrapper[4733]: I1204 19:20:17.335716 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:20:18 crc kubenswrapper[4733]: I1204 19:20:18.098398 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"353758af15d16464418cca9546b4d8b6ebe8b41cef4796d49135ac10e1a77f6b"} Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.356254 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.359052 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jnvbm" podUID="9ee5eee1-8eae-4de7-a2df-9ebbea833016" containerName="ovn-controller" probeResult="failure" output=< Dec 04 19:20:22 crc kubenswrapper[4733]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 19:20:22 crc kubenswrapper[4733]: > Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.373377 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-h6kk4" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.500151 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jnvbm-config-sz92r"] Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.501620 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.505004 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.519532 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jnvbm-config-sz92r"] Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.595601 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run-ovn\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.595663 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.595693 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-additional-scripts\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.595715 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-scripts\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.595733 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg66p\" (UniqueName: \"kubernetes.io/projected/b08183aa-2590-4146-9e2b-3c8acbbdd382-kube-api-access-pg66p\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.595761 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-log-ovn\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.696938 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run-ovn\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.697012 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.697038 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-additional-scripts\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.697057 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-scripts\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.697076 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg66p\" (UniqueName: \"kubernetes.io/projected/b08183aa-2590-4146-9e2b-3c8acbbdd382-kube-api-access-pg66p\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.697119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-log-ovn\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.697447 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-log-ovn\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.697472 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.697576 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run-ovn\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.698188 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-additional-scripts\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.699845 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-scripts\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.700260 4733 scope.go:117] "RemoveContainer" containerID="cb9de33098fddc8c46c84a918c754948d9f51cac41ce8e66745ee9dbd028e6f3" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.727315 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg66p\" (UniqueName: \"kubernetes.io/projected/b08183aa-2590-4146-9e2b-3c8acbbdd382-kube-api-access-pg66p\") pod \"ovn-controller-jnvbm-config-sz92r\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.734983 4733 scope.go:117] "RemoveContainer" containerID="a1a33d66584ca51460258f595aab008766474cb28995466693b8b8de9ba4a26c" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.830427 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.846896 4733 scope.go:117] "RemoveContainer" containerID="48847da5f73ece7aa31649bab70a49d9d756ea52b466ef1bfa18de1e9e821497" Dec 04 19:20:22 crc kubenswrapper[4733]: I1204 19:20:22.881032 4733 scope.go:117] "RemoveContainer" containerID="cd09e372d6385c95f93a88b440c3650f5bbd3b00f569ac6426f887332705565f" Dec 04 19:20:23 crc kubenswrapper[4733]: I1204 19:20:23.322013 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jnvbm-config-sz92r"] Dec 04 19:20:24 crc kubenswrapper[4733]: I1204 19:20:24.172892 4733 generic.go:334] "Generic (PLEG): container finished" podID="b08183aa-2590-4146-9e2b-3c8acbbdd382" containerID="e0bd19e089dde943d8b5f5cdbfbddb6f3d72e7dbbde25d33e22c346bf2448024" exitCode=0 Dec 04 19:20:24 crc kubenswrapper[4733]: I1204 19:20:24.172984 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jnvbm-config-sz92r" event={"ID":"b08183aa-2590-4146-9e2b-3c8acbbdd382","Type":"ContainerDied","Data":"e0bd19e089dde943d8b5f5cdbfbddb6f3d72e7dbbde25d33e22c346bf2448024"} Dec 04 19:20:24 crc kubenswrapper[4733]: I1204 19:20:24.174005 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jnvbm-config-sz92r" event={"ID":"b08183aa-2590-4146-9e2b-3c8acbbdd382","Type":"ContainerStarted","Data":"a2375251beaf7e33247e19421150addce20c62af34f2c9dc4bc64b612748b842"} Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.593407 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669223 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-additional-scripts\") pod \"b08183aa-2590-4146-9e2b-3c8acbbdd382\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669300 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-log-ovn\") pod \"b08183aa-2590-4146-9e2b-3c8acbbdd382\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669386 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run-ovn\") pod \"b08183aa-2590-4146-9e2b-3c8acbbdd382\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669433 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run\") pod \"b08183aa-2590-4146-9e2b-3c8acbbdd382\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669461 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg66p\" (UniqueName: \"kubernetes.io/projected/b08183aa-2590-4146-9e2b-3c8acbbdd382-kube-api-access-pg66p\") pod \"b08183aa-2590-4146-9e2b-3c8acbbdd382\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669481 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-scripts\") pod \"b08183aa-2590-4146-9e2b-3c8acbbdd382\" (UID: \"b08183aa-2590-4146-9e2b-3c8acbbdd382\") " Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669488 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b08183aa-2590-4146-9e2b-3c8acbbdd382" (UID: "b08183aa-2590-4146-9e2b-3c8acbbdd382"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669524 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run" (OuterVolumeSpecName: "var-run") pod "b08183aa-2590-4146-9e2b-3c8acbbdd382" (UID: "b08183aa-2590-4146-9e2b-3c8acbbdd382"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669532 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b08183aa-2590-4146-9e2b-3c8acbbdd382" (UID: "b08183aa-2590-4146-9e2b-3c8acbbdd382"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.669976 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b08183aa-2590-4146-9e2b-3c8acbbdd382" (UID: "b08183aa-2590-4146-9e2b-3c8acbbdd382"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.670007 4733 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.670021 4733 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.670030 4733 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b08183aa-2590-4146-9e2b-3c8acbbdd382-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.670464 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-scripts" (OuterVolumeSpecName: "scripts") pod "b08183aa-2590-4146-9e2b-3c8acbbdd382" (UID: "b08183aa-2590-4146-9e2b-3c8acbbdd382"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.674608 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b08183aa-2590-4146-9e2b-3c8acbbdd382-kube-api-access-pg66p" (OuterVolumeSpecName: "kube-api-access-pg66p") pod "b08183aa-2590-4146-9e2b-3c8acbbdd382" (UID: "b08183aa-2590-4146-9e2b-3c8acbbdd382"). InnerVolumeSpecName "kube-api-access-pg66p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.772245 4733 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.772288 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg66p\" (UniqueName: \"kubernetes.io/projected/b08183aa-2590-4146-9e2b-3c8acbbdd382-kube-api-access-pg66p\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:25 crc kubenswrapper[4733]: I1204 19:20:25.772302 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b08183aa-2590-4146-9e2b-3c8acbbdd382-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:26 crc kubenswrapper[4733]: I1204 19:20:26.204010 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jnvbm-config-sz92r" event={"ID":"b08183aa-2590-4146-9e2b-3c8acbbdd382","Type":"ContainerDied","Data":"a2375251beaf7e33247e19421150addce20c62af34f2c9dc4bc64b612748b842"} Dec 04 19:20:26 crc kubenswrapper[4733]: I1204 19:20:26.204331 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2375251beaf7e33247e19421150addce20c62af34f2c9dc4bc64b612748b842" Dec 04 19:20:26 crc kubenswrapper[4733]: I1204 19:20:26.204103 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jnvbm-config-sz92r" Dec 04 19:20:26 crc kubenswrapper[4733]: I1204 19:20:26.669887 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jnvbm-config-sz92r"] Dec 04 19:20:26 crc kubenswrapper[4733]: I1204 19:20:26.679969 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jnvbm-config-sz92r"] Dec 04 19:20:27 crc kubenswrapper[4733]: I1204 19:20:27.324120 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jnvbm" Dec 04 19:20:28 crc kubenswrapper[4733]: I1204 19:20:28.356575 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b08183aa-2590-4146-9e2b-3c8acbbdd382" path="/var/lib/kubelet/pods/b08183aa-2590-4146-9e2b-3c8acbbdd382/volumes" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.122521 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-xt2bg"] Dec 04 19:20:29 crc kubenswrapper[4733]: E1204 19:20:29.123498 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08183aa-2590-4146-9e2b-3c8acbbdd382" containerName="ovn-config" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.123522 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08183aa-2590-4146-9e2b-3c8acbbdd382" containerName="ovn-config" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.123773 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08183aa-2590-4146-9e2b-3c8acbbdd382" containerName="ovn-config" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.125025 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.126908 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.127151 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.127151 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.153858 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-xt2bg"] Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.248110 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8591138e-6ad7-491b-9e0d-45e6e50132de-hm-ports\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.248171 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8591138e-6ad7-491b-9e0d-45e6e50132de-config-data\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.248475 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8591138e-6ad7-491b-9e0d-45e6e50132de-scripts\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.248737 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8591138e-6ad7-491b-9e0d-45e6e50132de-config-data-merged\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.351189 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8591138e-6ad7-491b-9e0d-45e6e50132de-config-data-merged\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.351298 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8591138e-6ad7-491b-9e0d-45e6e50132de-hm-ports\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.351343 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8591138e-6ad7-491b-9e0d-45e6e50132de-config-data\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.351436 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8591138e-6ad7-491b-9e0d-45e6e50132de-scripts\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.351998 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8591138e-6ad7-491b-9e0d-45e6e50132de-config-data-merged\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.352329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8591138e-6ad7-491b-9e0d-45e6e50132de-hm-ports\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.365316 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8591138e-6ad7-491b-9e0d-45e6e50132de-config-data\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.366745 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8591138e-6ad7-491b-9e0d-45e6e50132de-scripts\") pod \"octavia-rsyslog-xt2bg\" (UID: \"8591138e-6ad7-491b-9e0d-45e6e50132de\") " pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.447250 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.757256 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-56c9f55b99-lhcrw"] Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.763121 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.769128 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.780971 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-56c9f55b99-lhcrw"] Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.862382 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43c750bd-8840-4acd-a479-871a6425b74d-httpd-config\") pod \"octavia-image-upload-56c9f55b99-lhcrw\" (UID: \"43c750bd-8840-4acd-a479-871a6425b74d\") " pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.862460 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43c750bd-8840-4acd-a479-871a6425b74d-amphora-image\") pod \"octavia-image-upload-56c9f55b99-lhcrw\" (UID: \"43c750bd-8840-4acd-a479-871a6425b74d\") " pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.964414 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43c750bd-8840-4acd-a479-871a6425b74d-httpd-config\") pod \"octavia-image-upload-56c9f55b99-lhcrw\" (UID: \"43c750bd-8840-4acd-a479-871a6425b74d\") " pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.964465 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43c750bd-8840-4acd-a479-871a6425b74d-amphora-image\") pod \"octavia-image-upload-56c9f55b99-lhcrw\" (UID: \"43c750bd-8840-4acd-a479-871a6425b74d\") " pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.964981 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43c750bd-8840-4acd-a479-871a6425b74d-amphora-image\") pod \"octavia-image-upload-56c9f55b99-lhcrw\" (UID: \"43c750bd-8840-4acd-a479-871a6425b74d\") " pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:20:29 crc kubenswrapper[4733]: I1204 19:20:29.972785 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43c750bd-8840-4acd-a479-871a6425b74d-httpd-config\") pod \"octavia-image-upload-56c9f55b99-lhcrw\" (UID: \"43c750bd-8840-4acd-a479-871a6425b74d\") " pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.025587 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-xt2bg"] Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.096825 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.239949 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-xt2bg" event={"ID":"8591138e-6ad7-491b-9e0d-45e6e50132de","Type":"ContainerStarted","Data":"be7fcc9a88d1632f74be0ad9eeb447d83665711c4e7b27fcad7979ca6ff097d2"} Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.442636 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-r2p6q"] Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.445927 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.447819 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.474376 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-r2p6q"] Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.569026 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-56c9f55b99-lhcrw"] Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.579917 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.580056 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-combined-ca-bundle\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.580167 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data-merged\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.580222 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-scripts\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.682951 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data-merged\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.683033 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-scripts\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.683127 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.683205 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-combined-ca-bundle\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.684544 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data-merged\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.690995 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-combined-ca-bundle\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.694184 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-scripts\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.695488 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data\") pod \"octavia-db-sync-r2p6q\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:30 crc kubenswrapper[4733]: I1204 19:20:30.769133 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:31 crc kubenswrapper[4733]: I1204 19:20:31.249666 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-r2p6q"] Dec 04 19:20:31 crc kubenswrapper[4733]: I1204 19:20:31.257331 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" event={"ID":"43c750bd-8840-4acd-a479-871a6425b74d","Type":"ContainerStarted","Data":"a1c1c154b9338e8c0ed558341c9e91753e5309aa041c4691995f50785d2fcbef"} Dec 04 19:20:32 crc kubenswrapper[4733]: I1204 19:20:32.272184 4733 generic.go:334] "Generic (PLEG): container finished" podID="f89bf105-5dad-4c4e-a810-d9cc9936cf48" containerID="6dc9a47033c60a204919e3488932469af34f67781518ae5452ec27729bfc6200" exitCode=0 Dec 04 19:20:32 crc kubenswrapper[4733]: I1204 19:20:32.272278 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r2p6q" event={"ID":"f89bf105-5dad-4c4e-a810-d9cc9936cf48","Type":"ContainerDied","Data":"6dc9a47033c60a204919e3488932469af34f67781518ae5452ec27729bfc6200"} Dec 04 19:20:32 crc kubenswrapper[4733]: I1204 19:20:32.272581 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r2p6q" event={"ID":"f89bf105-5dad-4c4e-a810-d9cc9936cf48","Type":"ContainerStarted","Data":"5bc5b1afcfbcdc5692ad0d92bd2d6764c44cc0e6ab571943be6a87e256263d77"} Dec 04 19:20:32 crc kubenswrapper[4733]: I1204 19:20:32.278068 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-xt2bg" event={"ID":"8591138e-6ad7-491b-9e0d-45e6e50132de","Type":"ContainerStarted","Data":"372ac1d2b2571db491d95476c9103dc33053fd3df6db209b68a067c00add84ae"} Dec 04 19:20:34 crc kubenswrapper[4733]: I1204 19:20:34.310875 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r2p6q" event={"ID":"f89bf105-5dad-4c4e-a810-d9cc9936cf48","Type":"ContainerStarted","Data":"95b64fb5660e9d6af7f652b88cd625bee0b60b8376b5f9d07e7db25bf6eca53d"} Dec 04 19:20:34 crc kubenswrapper[4733]: I1204 19:20:34.314513 4733 generic.go:334] "Generic (PLEG): container finished" podID="8591138e-6ad7-491b-9e0d-45e6e50132de" containerID="372ac1d2b2571db491d95476c9103dc33053fd3df6db209b68a067c00add84ae" exitCode=0 Dec 04 19:20:34 crc kubenswrapper[4733]: I1204 19:20:34.314770 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-xt2bg" event={"ID":"8591138e-6ad7-491b-9e0d-45e6e50132de","Type":"ContainerDied","Data":"372ac1d2b2571db491d95476c9103dc33053fd3df6db209b68a067c00add84ae"} Dec 04 19:20:34 crc kubenswrapper[4733]: I1204 19:20:34.330394 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-r2p6q" podStartSLOduration=4.330370207 podStartE2EDuration="4.330370207s" podCreationTimestamp="2025-12-04 19:20:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:20:34.328523626 +0000 UTC m=+6096.283884682" watchObservedRunningTime="2025-12-04 19:20:34.330370207 +0000 UTC m=+6096.285731263" Dec 04 19:20:37 crc kubenswrapper[4733]: I1204 19:20:37.904587 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:38 crc kubenswrapper[4733]: I1204 19:20:38.121907 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-59457b6897-wl54k" Dec 04 19:20:38 crc kubenswrapper[4733]: I1204 19:20:38.364083 4733 generic.go:334] "Generic (PLEG): container finished" podID="f89bf105-5dad-4c4e-a810-d9cc9936cf48" containerID="95b64fb5660e9d6af7f652b88cd625bee0b60b8376b5f9d07e7db25bf6eca53d" exitCode=0 Dec 04 19:20:38 crc kubenswrapper[4733]: I1204 19:20:38.364157 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r2p6q" event={"ID":"f89bf105-5dad-4c4e-a810-d9cc9936cf48","Type":"ContainerDied","Data":"95b64fb5660e9d6af7f652b88cd625bee0b60b8376b5f9d07e7db25bf6eca53d"} Dec 04 19:20:42 crc kubenswrapper[4733]: I1204 19:20:42.911627 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.036653 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data\") pod \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.036740 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-combined-ca-bundle\") pod \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.036804 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data-merged\") pod \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.036838 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-scripts\") pod \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\" (UID: \"f89bf105-5dad-4c4e-a810-d9cc9936cf48\") " Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.044026 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data" (OuterVolumeSpecName: "config-data") pod "f89bf105-5dad-4c4e-a810-d9cc9936cf48" (UID: "f89bf105-5dad-4c4e-a810-d9cc9936cf48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.058694 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-scripts" (OuterVolumeSpecName: "scripts") pod "f89bf105-5dad-4c4e-a810-d9cc9936cf48" (UID: "f89bf105-5dad-4c4e-a810-d9cc9936cf48"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.069068 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f89bf105-5dad-4c4e-a810-d9cc9936cf48" (UID: "f89bf105-5dad-4c4e-a810-d9cc9936cf48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.082631 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "f89bf105-5dad-4c4e-a810-d9cc9936cf48" (UID: "f89bf105-5dad-4c4e-a810-d9cc9936cf48"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.139093 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.139132 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.139143 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f89bf105-5dad-4c4e-a810-d9cc9936cf48-config-data-merged\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.139152 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f89bf105-5dad-4c4e-a810-d9cc9936cf48-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.420318 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r2p6q" event={"ID":"f89bf105-5dad-4c4e-a810-d9cc9936cf48","Type":"ContainerDied","Data":"5bc5b1afcfbcdc5692ad0d92bd2d6764c44cc0e6ab571943be6a87e256263d77"} Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.420652 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bc5b1afcfbcdc5692ad0d92bd2d6764c44cc0e6ab571943be6a87e256263d77" Dec 04 19:20:43 crc kubenswrapper[4733]: I1204 19:20:43.420388 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-r2p6q" Dec 04 19:20:44 crc kubenswrapper[4733]: I1204 19:20:44.432572 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-xt2bg" event={"ID":"8591138e-6ad7-491b-9e0d-45e6e50132de","Type":"ContainerStarted","Data":"43e4e859ddf70f88473bbf935782d2e07978cfcad51edfb16f2d96eb59584772"} Dec 04 19:20:44 crc kubenswrapper[4733]: I1204 19:20:44.433159 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:20:44 crc kubenswrapper[4733]: I1204 19:20:44.434469 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" event={"ID":"43c750bd-8840-4acd-a479-871a6425b74d","Type":"ContainerStarted","Data":"8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167"} Dec 04 19:20:44 crc kubenswrapper[4733]: I1204 19:20:44.458252 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-xt2bg" podStartSLOduration=2.2641631159999998 podStartE2EDuration="15.458228153s" podCreationTimestamp="2025-12-04 19:20:29 +0000 UTC" firstStartedPulling="2025-12-04 19:20:30.035179816 +0000 UTC m=+6091.990540862" lastFinishedPulling="2025-12-04 19:20:43.229244853 +0000 UTC m=+6105.184605899" observedRunningTime="2025-12-04 19:20:44.455600821 +0000 UTC m=+6106.410961867" watchObservedRunningTime="2025-12-04 19:20:44.458228153 +0000 UTC m=+6106.413589199" Dec 04 19:20:47 crc kubenswrapper[4733]: E1204 19:20:47.411591 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43c750bd_8840_4acd_a479_871a6425b74d.slice/crio-conmon-8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167.scope\": RecentStats: unable to find data in memory cache]" Dec 04 19:20:47 crc kubenswrapper[4733]: I1204 19:20:47.464325 4733 generic.go:334] "Generic (PLEG): container finished" podID="43c750bd-8840-4acd-a479-871a6425b74d" containerID="8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167" exitCode=0 Dec 04 19:20:47 crc kubenswrapper[4733]: I1204 19:20:47.464367 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" event={"ID":"43c750bd-8840-4acd-a479-871a6425b74d","Type":"ContainerDied","Data":"8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167"} Dec 04 19:20:48 crc kubenswrapper[4733]: I1204 19:20:48.479658 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" event={"ID":"43c750bd-8840-4acd-a479-871a6425b74d","Type":"ContainerStarted","Data":"42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876"} Dec 04 19:20:48 crc kubenswrapper[4733]: I1204 19:20:48.500450 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" podStartSLOduration=6.752849719 podStartE2EDuration="19.500428781s" podCreationTimestamp="2025-12-04 19:20:29 +0000 UTC" firstStartedPulling="2025-12-04 19:20:30.580199406 +0000 UTC m=+6092.535560452" lastFinishedPulling="2025-12-04 19:20:43.327778468 +0000 UTC m=+6105.283139514" observedRunningTime="2025-12-04 19:20:48.498523439 +0000 UTC m=+6110.453884505" watchObservedRunningTime="2025-12-04 19:20:48.500428781 +0000 UTC m=+6110.455789827" Dec 04 19:20:59 crc kubenswrapper[4733]: I1204 19:20:59.484197 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-xt2bg" Dec 04 19:21:08 crc kubenswrapper[4733]: I1204 19:21:08.870143 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-56c9f55b99-lhcrw"] Dec 04 19:21:08 crc kubenswrapper[4733]: I1204 19:21:08.871030 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" podUID="43c750bd-8840-4acd-a479-871a6425b74d" containerName="octavia-amphora-httpd" containerID="cri-o://42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876" gracePeriod=30 Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.523743 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.591473 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43c750bd-8840-4acd-a479-871a6425b74d-httpd-config\") pod \"43c750bd-8840-4acd-a479-871a6425b74d\" (UID: \"43c750bd-8840-4acd-a479-871a6425b74d\") " Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.591573 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43c750bd-8840-4acd-a479-871a6425b74d-amphora-image\") pod \"43c750bd-8840-4acd-a479-871a6425b74d\" (UID: \"43c750bd-8840-4acd-a479-871a6425b74d\") " Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.637934 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c750bd-8840-4acd-a479-871a6425b74d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "43c750bd-8840-4acd-a479-871a6425b74d" (UID: "43c750bd-8840-4acd-a479-871a6425b74d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.647381 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43c750bd-8840-4acd-a479-871a6425b74d-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "43c750bd-8840-4acd-a479-871a6425b74d" (UID: "43c750bd-8840-4acd-a479-871a6425b74d"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.694020 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43c750bd-8840-4acd-a479-871a6425b74d-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.694065 4733 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43c750bd-8840-4acd-a479-871a6425b74d-amphora-image\") on node \"crc\" DevicePath \"\"" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.714161 4733 generic.go:334] "Generic (PLEG): container finished" podID="43c750bd-8840-4acd-a479-871a6425b74d" containerID="42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876" exitCode=0 Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.714207 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" event={"ID":"43c750bd-8840-4acd-a479-871a6425b74d","Type":"ContainerDied","Data":"42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876"} Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.714270 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.714289 4733 scope.go:117] "RemoveContainer" containerID="42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.714277 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-56c9f55b99-lhcrw" event={"ID":"43c750bd-8840-4acd-a479-871a6425b74d","Type":"ContainerDied","Data":"a1c1c154b9338e8c0ed558341c9e91753e5309aa041c4691995f50785d2fcbef"} Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.754995 4733 scope.go:117] "RemoveContainer" containerID="8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.755759 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-56c9f55b99-lhcrw"] Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.764062 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-56c9f55b99-lhcrw"] Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.777594 4733 scope.go:117] "RemoveContainer" containerID="42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876" Dec 04 19:21:09 crc kubenswrapper[4733]: E1204 19:21:09.778045 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876\": container with ID starting with 42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876 not found: ID does not exist" containerID="42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.778093 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876"} err="failed to get container status \"42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876\": rpc error: code = NotFound desc = could not find container \"42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876\": container with ID starting with 42f4001fe3225cc6bc0ec8b429e9ad7cacfc2ee118f5bf1267828d439ec88876 not found: ID does not exist" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.778128 4733 scope.go:117] "RemoveContainer" containerID="8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167" Dec 04 19:21:09 crc kubenswrapper[4733]: E1204 19:21:09.778670 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167\": container with ID starting with 8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167 not found: ID does not exist" containerID="8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167" Dec 04 19:21:09 crc kubenswrapper[4733]: I1204 19:21:09.778709 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167"} err="failed to get container status \"8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167\": rpc error: code = NotFound desc = could not find container \"8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167\": container with ID starting with 8127bff1a119985eee228bddbd40d4d197bf3dda8e663c357d15b16a10ad1167 not found: ID does not exist" Dec 04 19:21:10 crc kubenswrapper[4733]: I1204 19:21:10.351897 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43c750bd-8840-4acd-a479-871a6425b74d" path="/var/lib/kubelet/pods/43c750bd-8840-4acd-a479-871a6425b74d/volumes" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.377610 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8c4p9"] Dec 04 19:21:27 crc kubenswrapper[4733]: E1204 19:21:27.378822 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c750bd-8840-4acd-a479-871a6425b74d" containerName="init" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.378839 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c750bd-8840-4acd-a479-871a6425b74d" containerName="init" Dec 04 19:21:27 crc kubenswrapper[4733]: E1204 19:21:27.378860 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bf105-5dad-4c4e-a810-d9cc9936cf48" containerName="init" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.378867 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bf105-5dad-4c4e-a810-d9cc9936cf48" containerName="init" Dec 04 19:21:27 crc kubenswrapper[4733]: E1204 19:21:27.378908 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c750bd-8840-4acd-a479-871a6425b74d" containerName="octavia-amphora-httpd" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.378918 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c750bd-8840-4acd-a479-871a6425b74d" containerName="octavia-amphora-httpd" Dec 04 19:21:27 crc kubenswrapper[4733]: E1204 19:21:27.378934 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bf105-5dad-4c4e-a810-d9cc9936cf48" containerName="octavia-db-sync" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.378942 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bf105-5dad-4c4e-a810-d9cc9936cf48" containerName="octavia-db-sync" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.379133 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c750bd-8840-4acd-a479-871a6425b74d" containerName="octavia-amphora-httpd" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.379163 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bf105-5dad-4c4e-a810-d9cc9936cf48" containerName="octavia-db-sync" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.380913 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.394965 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8c4p9"] Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.577810 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-catalog-content\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.577915 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-utilities\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.577986 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpn74\" (UniqueName: \"kubernetes.io/projected/3d9e051e-de89-480b-8755-3e3f248e57da-kube-api-access-dpn74\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.679965 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-catalog-content\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.680097 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-utilities\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.680167 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpn74\" (UniqueName: \"kubernetes.io/projected/3d9e051e-de89-480b-8755-3e3f248e57da-kube-api-access-dpn74\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.680664 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-utilities\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.680784 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-catalog-content\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:27 crc kubenswrapper[4733]: I1204 19:21:27.718116 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpn74\" (UniqueName: \"kubernetes.io/projected/3d9e051e-de89-480b-8755-3e3f248e57da-kube-api-access-dpn74\") pod \"redhat-operators-8c4p9\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:28 crc kubenswrapper[4733]: I1204 19:21:28.002172 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:28 crc kubenswrapper[4733]: I1204 19:21:28.481978 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8c4p9"] Dec 04 19:21:29 crc kubenswrapper[4733]: I1204 19:21:29.151076 4733 generic.go:334] "Generic (PLEG): container finished" podID="3d9e051e-de89-480b-8755-3e3f248e57da" containerID="82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6" exitCode=0 Dec 04 19:21:29 crc kubenswrapper[4733]: I1204 19:21:29.151158 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c4p9" event={"ID":"3d9e051e-de89-480b-8755-3e3f248e57da","Type":"ContainerDied","Data":"82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6"} Dec 04 19:21:29 crc kubenswrapper[4733]: I1204 19:21:29.151415 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c4p9" event={"ID":"3d9e051e-de89-480b-8755-3e3f248e57da","Type":"ContainerStarted","Data":"1f8a55c5656e5e92f4bdbb7880cc0fcad5bba235210dd487f4fe62fae73ad1dd"} Dec 04 19:21:30 crc kubenswrapper[4733]: I1204 19:21:30.164606 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c4p9" event={"ID":"3d9e051e-de89-480b-8755-3e3f248e57da","Type":"ContainerStarted","Data":"b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f"} Dec 04 19:21:32 crc kubenswrapper[4733]: I1204 19:21:32.186245 4733 generic.go:334] "Generic (PLEG): container finished" podID="3d9e051e-de89-480b-8755-3e3f248e57da" containerID="b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f" exitCode=0 Dec 04 19:21:32 crc kubenswrapper[4733]: I1204 19:21:32.186384 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c4p9" event={"ID":"3d9e051e-de89-480b-8755-3e3f248e57da","Type":"ContainerDied","Data":"b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f"} Dec 04 19:21:33 crc kubenswrapper[4733]: I1204 19:21:33.197207 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c4p9" event={"ID":"3d9e051e-de89-480b-8755-3e3f248e57da","Type":"ContainerStarted","Data":"55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc"} Dec 04 19:21:33 crc kubenswrapper[4733]: I1204 19:21:33.219056 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8c4p9" podStartSLOduration=2.660117507 podStartE2EDuration="6.21903137s" podCreationTimestamp="2025-12-04 19:21:27 +0000 UTC" firstStartedPulling="2025-12-04 19:21:29.153253181 +0000 UTC m=+6151.108614227" lastFinishedPulling="2025-12-04 19:21:32.712167044 +0000 UTC m=+6154.667528090" observedRunningTime="2025-12-04 19:21:33.214093206 +0000 UTC m=+6155.169454242" watchObservedRunningTime="2025-12-04 19:21:33.21903137 +0000 UTC m=+6155.174392416" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.477723 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-mj4sc"] Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.479740 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.484054 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.484153 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.484510 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.498118 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-mj4sc"] Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.612305 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-amphora-certs\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.612409 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-combined-ca-bundle\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.612443 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5114b53c-38d9-45b5-9f9b-50713a9f8abd-config-data-merged\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.612635 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-scripts\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.613018 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-config-data\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.613151 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/5114b53c-38d9-45b5-9f9b-50713a9f8abd-hm-ports\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.714749 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-scripts\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.714892 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-config-data\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.714930 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/5114b53c-38d9-45b5-9f9b-50713a9f8abd-hm-ports\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.714971 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-amphora-certs\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.715038 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-combined-ca-bundle\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.715069 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5114b53c-38d9-45b5-9f9b-50713a9f8abd-config-data-merged\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.715590 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5114b53c-38d9-45b5-9f9b-50713a9f8abd-config-data-merged\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.716087 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/5114b53c-38d9-45b5-9f9b-50713a9f8abd-hm-ports\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.722176 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-scripts\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.724900 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-combined-ca-bundle\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.724929 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-amphora-certs\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.736532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5114b53c-38d9-45b5-9f9b-50713a9f8abd-config-data\") pod \"octavia-healthmanager-mj4sc\" (UID: \"5114b53c-38d9-45b5-9f9b-50713a9f8abd\") " pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:34 crc kubenswrapper[4733]: I1204 19:21:34.810204 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.231304 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-85ctw"] Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.233613 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.240170 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.240219 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.242173 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-85ctw"] Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.331319 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-combined-ca-bundle\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.331770 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-scripts\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.331834 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2d9abffe-6c27-4cf4-b943-3ad0f694228f-config-data-merged\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.331952 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-config-data\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.332031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2d9abffe-6c27-4cf4-b943-3ad0f694228f-hm-ports\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.332082 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-amphora-certs\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: W1204 19:21:35.352597 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5114b53c_38d9_45b5_9f9b_50713a9f8abd.slice/crio-42e5713fb82c7deb91c72cd4c15841e8688d790ad426f13bcdb869eb80b2a2bd WatchSource:0}: Error finding container 42e5713fb82c7deb91c72cd4c15841e8688d790ad426f13bcdb869eb80b2a2bd: Status 404 returned error can't find the container with id 42e5713fb82c7deb91c72cd4c15841e8688d790ad426f13bcdb869eb80b2a2bd Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.352735 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-mj4sc"] Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.446108 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-combined-ca-bundle\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.446201 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-scripts\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.446242 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2d9abffe-6c27-4cf4-b943-3ad0f694228f-config-data-merged\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.446317 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-config-data\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.446379 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2d9abffe-6c27-4cf4-b943-3ad0f694228f-hm-ports\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.446425 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-amphora-certs\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.448202 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2d9abffe-6c27-4cf4-b943-3ad0f694228f-config-data-merged\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.449578 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2d9abffe-6c27-4cf4-b943-3ad0f694228f-hm-ports\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.456720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-amphora-certs\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.457496 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-combined-ca-bundle\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.463156 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-config-data\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.483954 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d9abffe-6c27-4cf4-b943-3ad0f694228f-scripts\") pod \"octavia-housekeeping-85ctw\" (UID: \"2d9abffe-6c27-4cf4-b943-3ad0f694228f\") " pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:35 crc kubenswrapper[4733]: I1204 19:21:35.556318 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:36 crc kubenswrapper[4733]: I1204 19:21:36.183513 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-85ctw"] Dec 04 19:21:36 crc kubenswrapper[4733]: W1204 19:21:36.192899 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d9abffe_6c27_4cf4_b943_3ad0f694228f.slice/crio-7ae0f0f1da955584b942b6f34e2ec25291a016532ef98a1938a2ed61f9599e67 WatchSource:0}: Error finding container 7ae0f0f1da955584b942b6f34e2ec25291a016532ef98a1938a2ed61f9599e67: Status 404 returned error can't find the container with id 7ae0f0f1da955584b942b6f34e2ec25291a016532ef98a1938a2ed61f9599e67 Dec 04 19:21:36 crc kubenswrapper[4733]: I1204 19:21:36.238719 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-85ctw" event={"ID":"2d9abffe-6c27-4cf4-b943-3ad0f694228f","Type":"ContainerStarted","Data":"7ae0f0f1da955584b942b6f34e2ec25291a016532ef98a1938a2ed61f9599e67"} Dec 04 19:21:36 crc kubenswrapper[4733]: I1204 19:21:36.239663 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-mj4sc" event={"ID":"5114b53c-38d9-45b5-9f9b-50713a9f8abd","Type":"ContainerStarted","Data":"42e5713fb82c7deb91c72cd4c15841e8688d790ad426f13bcdb869eb80b2a2bd"} Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.086577 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-x47tj"] Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.088448 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.091600 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.091807 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.129720 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-x47tj"] Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.255757 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-mj4sc" event={"ID":"5114b53c-38d9-45b5-9f9b-50713a9f8abd","Type":"ContainerStarted","Data":"a30d72e1aec20f9259d846e42e0d01b8c4399dcc8e26f9d7fe68d9272ec04107"} Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.284003 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-combined-ca-bundle\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.285097 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4ba8c639-2dcf-48f5-a004-0f5e16aee627-config-data-merged\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.285209 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-config-data\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.285308 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4ba8c639-2dcf-48f5-a004-0f5e16aee627-hm-ports\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.285426 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-amphora-certs\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.285541 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-scripts\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.387756 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4ba8c639-2dcf-48f5-a004-0f5e16aee627-hm-ports\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.388172 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-amphora-certs\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.388252 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-scripts\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.388315 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-combined-ca-bundle\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.388440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4ba8c639-2dcf-48f5-a004-0f5e16aee627-config-data-merged\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.388479 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-config-data\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.389268 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4ba8c639-2dcf-48f5-a004-0f5e16aee627-hm-ports\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.391901 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4ba8c639-2dcf-48f5-a004-0f5e16aee627-config-data-merged\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.394834 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-scripts\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.395425 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-config-data\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.396394 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-combined-ca-bundle\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.397149 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/4ba8c639-2dcf-48f5-a004-0f5e16aee627-amphora-certs\") pod \"octavia-worker-x47tj\" (UID: \"4ba8c639-2dcf-48f5-a004-0f5e16aee627\") " pod="openstack/octavia-worker-x47tj" Dec 04 19:21:37 crc kubenswrapper[4733]: I1204 19:21:37.412838 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-x47tj" Dec 04 19:21:38 crc kubenswrapper[4733]: I1204 19:21:38.002851 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:38 crc kubenswrapper[4733]: I1204 19:21:38.003234 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:38 crc kubenswrapper[4733]: I1204 19:21:38.515459 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-x47tj"] Dec 04 19:21:38 crc kubenswrapper[4733]: W1204 19:21:38.516902 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ba8c639_2dcf_48f5_a004_0f5e16aee627.slice/crio-462e0c6bd6d6efda0e100d412cdd4c61ecf0b3a9eaf6646941cb307f75e4d427 WatchSource:0}: Error finding container 462e0c6bd6d6efda0e100d412cdd4c61ecf0b3a9eaf6646941cb307f75e4d427: Status 404 returned error can't find the container with id 462e0c6bd6d6efda0e100d412cdd4c61ecf0b3a9eaf6646941cb307f75e4d427 Dec 04 19:21:38 crc kubenswrapper[4733]: E1204 19:21:38.757979 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5114b53c_38d9_45b5_9f9b_50713a9f8abd.slice/crio-a30d72e1aec20f9259d846e42e0d01b8c4399dcc8e26f9d7fe68d9272ec04107.scope\": RecentStats: unable to find data in memory cache]" Dec 04 19:21:39 crc kubenswrapper[4733]: I1204 19:21:39.146868 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8c4p9" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="registry-server" probeResult="failure" output=< Dec 04 19:21:39 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 19:21:39 crc kubenswrapper[4733]: > Dec 04 19:21:39 crc kubenswrapper[4733]: I1204 19:21:39.274534 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-x47tj" event={"ID":"4ba8c639-2dcf-48f5-a004-0f5e16aee627","Type":"ContainerStarted","Data":"462e0c6bd6d6efda0e100d412cdd4c61ecf0b3a9eaf6646941cb307f75e4d427"} Dec 04 19:21:39 crc kubenswrapper[4733]: I1204 19:21:39.286905 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-85ctw" event={"ID":"2d9abffe-6c27-4cf4-b943-3ad0f694228f","Type":"ContainerStarted","Data":"14ccc0975cd79b163b5afe0c3f0583e64a73859b61994a44c1cee721dd9fb7f5"} Dec 04 19:21:39 crc kubenswrapper[4733]: I1204 19:21:39.290156 4733 generic.go:334] "Generic (PLEG): container finished" podID="5114b53c-38d9-45b5-9f9b-50713a9f8abd" containerID="a30d72e1aec20f9259d846e42e0d01b8c4399dcc8e26f9d7fe68d9272ec04107" exitCode=0 Dec 04 19:21:39 crc kubenswrapper[4733]: I1204 19:21:39.290202 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-mj4sc" event={"ID":"5114b53c-38d9-45b5-9f9b-50713a9f8abd","Type":"ContainerDied","Data":"a30d72e1aec20f9259d846e42e0d01b8c4399dcc8e26f9d7fe68d9272ec04107"} Dec 04 19:21:40 crc kubenswrapper[4733]: I1204 19:21:40.303618 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d9abffe-6c27-4cf4-b943-3ad0f694228f" containerID="14ccc0975cd79b163b5afe0c3f0583e64a73859b61994a44c1cee721dd9fb7f5" exitCode=0 Dec 04 19:21:40 crc kubenswrapper[4733]: I1204 19:21:40.303805 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-85ctw" event={"ID":"2d9abffe-6c27-4cf4-b943-3ad0f694228f","Type":"ContainerDied","Data":"14ccc0975cd79b163b5afe0c3f0583e64a73859b61994a44c1cee721dd9fb7f5"} Dec 04 19:21:40 crc kubenswrapper[4733]: I1204 19:21:40.314360 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-mj4sc" event={"ID":"5114b53c-38d9-45b5-9f9b-50713a9f8abd","Type":"ContainerStarted","Data":"1554ea384552b375517190eb6d3463d8232cad468f2e7d6ecfe617da52dfa13f"} Dec 04 19:21:40 crc kubenswrapper[4733]: I1204 19:21:40.314687 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:40 crc kubenswrapper[4733]: I1204 19:21:40.355820 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-mj4sc" podStartSLOduration=6.35578135 podStartE2EDuration="6.35578135s" podCreationTimestamp="2025-12-04 19:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:21:40.342898031 +0000 UTC m=+6162.298259087" watchObservedRunningTime="2025-12-04 19:21:40.35578135 +0000 UTC m=+6162.311142396" Dec 04 19:21:42 crc kubenswrapper[4733]: I1204 19:21:42.355068 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-x47tj" event={"ID":"4ba8c639-2dcf-48f5-a004-0f5e16aee627","Type":"ContainerStarted","Data":"77addadf024704e1b89bc56151e1d463df758a0e9b169aed4ed1446aae795c98"} Dec 04 19:21:42 crc kubenswrapper[4733]: I1204 19:21:42.356087 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:42 crc kubenswrapper[4733]: I1204 19:21:42.356112 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-85ctw" event={"ID":"2d9abffe-6c27-4cf4-b943-3ad0f694228f","Type":"ContainerStarted","Data":"f59ea0d3e20314641adc76e870a596f25add5336b3ae313cfc3073bf43fc49f6"} Dec 04 19:21:42 crc kubenswrapper[4733]: I1204 19:21:42.404072 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-85ctw" podStartSLOduration=5.761396675 podStartE2EDuration="7.404047799s" podCreationTimestamp="2025-12-04 19:21:35 +0000 UTC" firstStartedPulling="2025-12-04 19:21:36.19639681 +0000 UTC m=+6158.151757856" lastFinishedPulling="2025-12-04 19:21:37.839047934 +0000 UTC m=+6159.794408980" observedRunningTime="2025-12-04 19:21:42.399095575 +0000 UTC m=+6164.354456641" watchObservedRunningTime="2025-12-04 19:21:42.404047799 +0000 UTC m=+6164.359408875" Dec 04 19:21:43 crc kubenswrapper[4733]: I1204 19:21:43.366892 4733 generic.go:334] "Generic (PLEG): container finished" podID="4ba8c639-2dcf-48f5-a004-0f5e16aee627" containerID="77addadf024704e1b89bc56151e1d463df758a0e9b169aed4ed1446aae795c98" exitCode=0 Dec 04 19:21:43 crc kubenswrapper[4733]: I1204 19:21:43.367322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-x47tj" event={"ID":"4ba8c639-2dcf-48f5-a004-0f5e16aee627","Type":"ContainerDied","Data":"77addadf024704e1b89bc56151e1d463df758a0e9b169aed4ed1446aae795c98"} Dec 04 19:21:44 crc kubenswrapper[4733]: I1204 19:21:44.389631 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-x47tj" event={"ID":"4ba8c639-2dcf-48f5-a004-0f5e16aee627","Type":"ContainerStarted","Data":"c6d8e0f2bfede7be95aedf6bc266cb30d3a4b585407bc17fe27b250654b14d65"} Dec 04 19:21:44 crc kubenswrapper[4733]: I1204 19:21:44.390418 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-x47tj" Dec 04 19:21:44 crc kubenswrapper[4733]: I1204 19:21:44.417678 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-x47tj" podStartSLOduration=4.933815995 podStartE2EDuration="7.41765945s" podCreationTimestamp="2025-12-04 19:21:37 +0000 UTC" firstStartedPulling="2025-12-04 19:21:38.519214334 +0000 UTC m=+6160.474575380" lastFinishedPulling="2025-12-04 19:21:41.003057779 +0000 UTC m=+6162.958418835" observedRunningTime="2025-12-04 19:21:44.409466048 +0000 UTC m=+6166.364827104" watchObservedRunningTime="2025-12-04 19:21:44.41765945 +0000 UTC m=+6166.373020496" Dec 04 19:21:48 crc kubenswrapper[4733]: I1204 19:21:48.057974 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:48 crc kubenswrapper[4733]: I1204 19:21:48.114435 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:48 crc kubenswrapper[4733]: I1204 19:21:48.300164 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8c4p9"] Dec 04 19:21:49 crc kubenswrapper[4733]: I1204 19:21:49.441139 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8c4p9" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="registry-server" containerID="cri-o://55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc" gracePeriod=2 Dec 04 19:21:49 crc kubenswrapper[4733]: I1204 19:21:49.843916 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-mj4sc" Dec 04 19:21:49 crc kubenswrapper[4733]: I1204 19:21:49.918586 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:49 crc kubenswrapper[4733]: I1204 19:21:49.953508 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpn74\" (UniqueName: \"kubernetes.io/projected/3d9e051e-de89-480b-8755-3e3f248e57da-kube-api-access-dpn74\") pod \"3d9e051e-de89-480b-8755-3e3f248e57da\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " Dec 04 19:21:49 crc kubenswrapper[4733]: I1204 19:21:49.953573 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-utilities\") pod \"3d9e051e-de89-480b-8755-3e3f248e57da\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " Dec 04 19:21:49 crc kubenswrapper[4733]: I1204 19:21:49.953868 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-catalog-content\") pod \"3d9e051e-de89-480b-8755-3e3f248e57da\" (UID: \"3d9e051e-de89-480b-8755-3e3f248e57da\") " Dec 04 19:21:49 crc kubenswrapper[4733]: I1204 19:21:49.957059 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-utilities" (OuterVolumeSpecName: "utilities") pod "3d9e051e-de89-480b-8755-3e3f248e57da" (UID: "3d9e051e-de89-480b-8755-3e3f248e57da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:21:49 crc kubenswrapper[4733]: I1204 19:21:49.960015 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9e051e-de89-480b-8755-3e3f248e57da-kube-api-access-dpn74" (OuterVolumeSpecName: "kube-api-access-dpn74") pod "3d9e051e-de89-480b-8755-3e3f248e57da" (UID: "3d9e051e-de89-480b-8755-3e3f248e57da"). InnerVolumeSpecName "kube-api-access-dpn74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.056299 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpn74\" (UniqueName: \"kubernetes.io/projected/3d9e051e-de89-480b-8755-3e3f248e57da-kube-api-access-dpn74\") on node \"crc\" DevicePath \"\"" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.056343 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.115963 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d9e051e-de89-480b-8755-3e3f248e57da" (UID: "3d9e051e-de89-480b-8755-3e3f248e57da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.158825 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9e051e-de89-480b-8755-3e3f248e57da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.453389 4733 generic.go:334] "Generic (PLEG): container finished" podID="3d9e051e-de89-480b-8755-3e3f248e57da" containerID="55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc" exitCode=0 Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.453438 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c4p9" event={"ID":"3d9e051e-de89-480b-8755-3e3f248e57da","Type":"ContainerDied","Data":"55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc"} Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.453469 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c4p9" event={"ID":"3d9e051e-de89-480b-8755-3e3f248e57da","Type":"ContainerDied","Data":"1f8a55c5656e5e92f4bdbb7880cc0fcad5bba235210dd487f4fe62fae73ad1dd"} Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.453491 4733 scope.go:117] "RemoveContainer" containerID="55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.453639 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8c4p9" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.481593 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8c4p9"] Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.484335 4733 scope.go:117] "RemoveContainer" containerID="b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.490872 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8c4p9"] Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.510177 4733 scope.go:117] "RemoveContainer" containerID="82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.558330 4733 scope.go:117] "RemoveContainer" containerID="55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc" Dec 04 19:21:50 crc kubenswrapper[4733]: E1204 19:21:50.558748 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc\": container with ID starting with 55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc not found: ID does not exist" containerID="55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.558826 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc"} err="failed to get container status \"55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc\": rpc error: code = NotFound desc = could not find container \"55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc\": container with ID starting with 55fdbdf821b74caf02b082deee3d7ceb74049184221caadbcfc50ba840271fdc not found: ID does not exist" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.558855 4733 scope.go:117] "RemoveContainer" containerID="b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f" Dec 04 19:21:50 crc kubenswrapper[4733]: E1204 19:21:50.559256 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f\": container with ID starting with b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f not found: ID does not exist" containerID="b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.559289 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f"} err="failed to get container status \"b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f\": rpc error: code = NotFound desc = could not find container \"b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f\": container with ID starting with b3dfb766231aee5f4f034e87ace7b95ba49caa2cf80ea9033accd2e70359390f not found: ID does not exist" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.559310 4733 scope.go:117] "RemoveContainer" containerID="82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6" Dec 04 19:21:50 crc kubenswrapper[4733]: E1204 19:21:50.559665 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6\": container with ID starting with 82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6 not found: ID does not exist" containerID="82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.559704 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6"} err="failed to get container status \"82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6\": rpc error: code = NotFound desc = could not find container \"82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6\": container with ID starting with 82ee6c4e928ee9bfca338b1c37ad4b5324e0643340e947d8fc86ea354775d9c6 not found: ID does not exist" Dec 04 19:21:50 crc kubenswrapper[4733]: I1204 19:21:50.586752 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-85ctw" Dec 04 19:21:52 crc kubenswrapper[4733]: I1204 19:21:52.351387 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" path="/var/lib/kubelet/pods/3d9e051e-de89-480b-8755-3e3f248e57da/volumes" Dec 04 19:21:52 crc kubenswrapper[4733]: I1204 19:21:52.446736 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-x47tj" Dec 04 19:22:23 crc kubenswrapper[4733]: I1204 19:22:23.177211 4733 scope.go:117] "RemoveContainer" containerID="e2f457072bddbd8e832eb075ca553f01885aa55a715bbbaf7d273270a587d110" Dec 04 19:22:23 crc kubenswrapper[4733]: I1204 19:22:23.206849 4733 scope.go:117] "RemoveContainer" containerID="88850cf6b8ced264c71663128de23318f3409d85d1059c4234ac39ea79e79932" Dec 04 19:22:30 crc kubenswrapper[4733]: I1204 19:22:30.040894 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-983b-account-create-update-gr7xd"] Dec 04 19:22:30 crc kubenswrapper[4733]: I1204 19:22:30.051716 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-7s2bl"] Dec 04 19:22:30 crc kubenswrapper[4733]: I1204 19:22:30.062374 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-983b-account-create-update-gr7xd"] Dec 04 19:22:30 crc kubenswrapper[4733]: I1204 19:22:30.071089 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-7s2bl"] Dec 04 19:22:30 crc kubenswrapper[4733]: I1204 19:22:30.347473 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40f10d39-9d06-4362-b9f8-d4e05f623cdc" path="/var/lib/kubelet/pods/40f10d39-9d06-4362-b9f8-d4e05f623cdc/volumes" Dec 04 19:22:30 crc kubenswrapper[4733]: I1204 19:22:30.348173 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f7104c6-4073-40e7-b906-a659739adbb1" path="/var/lib/kubelet/pods/9f7104c6-4073-40e7-b906-a659739adbb1/volumes" Dec 04 19:22:36 crc kubenswrapper[4733]: I1204 19:22:36.038196 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-2pcqw"] Dec 04 19:22:36 crc kubenswrapper[4733]: I1204 19:22:36.048581 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-2pcqw"] Dec 04 19:22:36 crc kubenswrapper[4733]: I1204 19:22:36.349094 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57823d82-7ca6-48c9-8dbf-2f8f3360919d" path="/var/lib/kubelet/pods/57823d82-7ca6-48c9-8dbf-2f8f3360919d/volumes" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.725174 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-9d8bc57f7-p95j8"] Dec 04 19:22:44 crc kubenswrapper[4733]: E1204 19:22:44.726047 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="extract-utilities" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.726059 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="extract-utilities" Dec 04 19:22:44 crc kubenswrapper[4733]: E1204 19:22:44.726075 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="extract-content" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.726083 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="extract-content" Dec 04 19:22:44 crc kubenswrapper[4733]: E1204 19:22:44.726107 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="registry-server" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.726113 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="registry-server" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.726300 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9e051e-de89-480b-8755-3e3f248e57da" containerName="registry-server" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.727235 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.729568 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.729721 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.730881 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-r75kp" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.731899 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.781545 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9d8bc57f7-p95j8"] Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.815243 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.815505 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerName="glance-log" containerID="cri-o://3998038ce1e96df93e35f5935b8e0f2681e565a155bc670d2d3a88053dbec451" gracePeriod=30 Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.815874 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerName="glance-httpd" containerID="cri-o://f31aa29f94996c0ece09b0fa003c32cb17412c5b29b02e1ac58ae7b31cf6a428" gracePeriod=30 Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.865449 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-scripts\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.865575 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-config-data\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.865610 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d42413-6ae0-4845-b25c-4d19058854f5-logs\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.865684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqlnz\" (UniqueName: \"kubernetes.io/projected/66d42413-6ae0-4845-b25c-4d19058854f5-kube-api-access-sqlnz\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.865714 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66d42413-6ae0-4845-b25c-4d19058854f5-horizon-secret-key\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.865858 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.866066 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerName="glance-log" containerID="cri-o://de21d2bb957d01d50a27342f3b90c33caf498e96ba91ca2639f26e7c0dccc6c4" gracePeriod=30 Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.866576 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerName="glance-httpd" containerID="cri-o://1e0372feada67ec46db547f1eb5f49dd406f57ccd058b6aefa078af1a4c96559" gracePeriod=30 Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.877846 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-989dbf8fc-q2crf"] Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.879529 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.888471 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-989dbf8fc-q2crf"] Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.967118 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqlnz\" (UniqueName: \"kubernetes.io/projected/66d42413-6ae0-4845-b25c-4d19058854f5-kube-api-access-sqlnz\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.967180 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66d42413-6ae0-4845-b25c-4d19058854f5-horizon-secret-key\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.967250 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-scripts\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.967341 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-config-data\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.967395 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d42413-6ae0-4845-b25c-4d19058854f5-logs\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.968375 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d42413-6ae0-4845-b25c-4d19058854f5-logs\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.969620 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-scripts\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.970291 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-config-data\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.974747 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66d42413-6ae0-4845-b25c-4d19058854f5-horizon-secret-key\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:44 crc kubenswrapper[4733]: I1204 19:22:44.986395 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqlnz\" (UniqueName: \"kubernetes.io/projected/66d42413-6ae0-4845-b25c-4d19058854f5-kube-api-access-sqlnz\") pod \"horizon-9d8bc57f7-p95j8\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.056913 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.069613 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/648988d6-2223-4c06-b39f-c899a2778082-horizon-secret-key\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.069697 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-config-data\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.069726 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmxw4\" (UniqueName: \"kubernetes.io/projected/648988d6-2223-4c06-b39f-c899a2778082-kube-api-access-zmxw4\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.069917 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-scripts\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.070051 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/648988d6-2223-4c06-b39f-c899a2778082-logs\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.173335 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-scripts\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.173411 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/648988d6-2223-4c06-b39f-c899a2778082-logs\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.173474 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/648988d6-2223-4c06-b39f-c899a2778082-horizon-secret-key\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.173524 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-config-data\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.173580 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmxw4\" (UniqueName: \"kubernetes.io/projected/648988d6-2223-4c06-b39f-c899a2778082-kube-api-access-zmxw4\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.173884 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/648988d6-2223-4c06-b39f-c899a2778082-logs\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.174717 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-config-data\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.175175 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-scripts\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.178302 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/648988d6-2223-4c06-b39f-c899a2778082-horizon-secret-key\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.198447 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmxw4\" (UniqueName: \"kubernetes.io/projected/648988d6-2223-4c06-b39f-c899a2778082-kube-api-access-zmxw4\") pod \"horizon-989dbf8fc-q2crf\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.214610 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.362760 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.363147 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.513931 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-989dbf8fc-q2crf"] Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.550998 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66d7c5c989-8t52p"] Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.552974 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.571311 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66d7c5c989-8t52p"] Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.686824 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f339b4d-9773-4b73-8828-7cbe31882ff7-horizon-secret-key\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.686884 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-config-data\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.686960 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f339b4d-9773-4b73-8828-7cbe31882ff7-logs\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.687073 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-scripts\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.687101 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6w66\" (UniqueName: \"kubernetes.io/projected/8f339b4d-9773-4b73-8828-7cbe31882ff7-kube-api-access-n6w66\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.789009 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f339b4d-9773-4b73-8828-7cbe31882ff7-logs\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.789171 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-scripts\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.789211 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6w66\" (UniqueName: \"kubernetes.io/projected/8f339b4d-9773-4b73-8828-7cbe31882ff7-kube-api-access-n6w66\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.789239 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f339b4d-9773-4b73-8828-7cbe31882ff7-horizon-secret-key\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.789281 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-config-data\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.790084 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-scripts\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.791101 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f339b4d-9773-4b73-8828-7cbe31882ff7-logs\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.791662 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-config-data\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.798020 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f339b4d-9773-4b73-8828-7cbe31882ff7-horizon-secret-key\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.806296 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9d8bc57f7-p95j8"] Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.809489 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6w66\" (UniqueName: \"kubernetes.io/projected/8f339b4d-9773-4b73-8828-7cbe31882ff7-kube-api-access-n6w66\") pod \"horizon-66d7c5c989-8t52p\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.872100 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-989dbf8fc-q2crf"] Dec 04 19:22:45 crc kubenswrapper[4733]: I1204 19:22:45.893564 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:46 crc kubenswrapper[4733]: I1204 19:22:46.025513 4733 generic.go:334] "Generic (PLEG): container finished" podID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerID="de21d2bb957d01d50a27342f3b90c33caf498e96ba91ca2639f26e7c0dccc6c4" exitCode=143 Dec 04 19:22:46 crc kubenswrapper[4733]: I1204 19:22:46.025607 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdf21970-8b52-4c07-81aa-d786baa2bb58","Type":"ContainerDied","Data":"de21d2bb957d01d50a27342f3b90c33caf498e96ba91ca2639f26e7c0dccc6c4"} Dec 04 19:22:46 crc kubenswrapper[4733]: I1204 19:22:46.028724 4733 generic.go:334] "Generic (PLEG): container finished" podID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerID="3998038ce1e96df93e35f5935b8e0f2681e565a155bc670d2d3a88053dbec451" exitCode=143 Dec 04 19:22:46 crc kubenswrapper[4733]: I1204 19:22:46.028744 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a09af527-a663-4dfc-89f0-3e0d3f41c048","Type":"ContainerDied","Data":"3998038ce1e96df93e35f5935b8e0f2681e565a155bc670d2d3a88053dbec451"} Dec 04 19:22:46 crc kubenswrapper[4733]: I1204 19:22:46.030713 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d8bc57f7-p95j8" event={"ID":"66d42413-6ae0-4845-b25c-4d19058854f5","Type":"ContainerStarted","Data":"c09e972ecce60889651da09566a2b851d1ade38412fb02ab7d89f5c106af1e3b"} Dec 04 19:22:46 crc kubenswrapper[4733]: I1204 19:22:46.032530 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-989dbf8fc-q2crf" event={"ID":"648988d6-2223-4c06-b39f-c899a2778082","Type":"ContainerStarted","Data":"0d51de6cb0e2e9c8f29954909c3f23ded6d96dd3c7133ec9fdcdce9096583ac6"} Dec 04 19:22:46 crc kubenswrapper[4733]: I1204 19:22:46.348437 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66d7c5c989-8t52p"] Dec 04 19:22:46 crc kubenswrapper[4733]: W1204 19:22:46.355747 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f339b4d_9773_4b73_8828_7cbe31882ff7.slice/crio-fe38323aa3e661b6285e056850d733fecd56bf524b0c2ac282c17e48d5c8ace5 WatchSource:0}: Error finding container fe38323aa3e661b6285e056850d733fecd56bf524b0c2ac282c17e48d5c8ace5: Status 404 returned error can't find the container with id fe38323aa3e661b6285e056850d733fecd56bf524b0c2ac282c17e48d5c8ace5 Dec 04 19:22:47 crc kubenswrapper[4733]: I1204 19:22:47.048873 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d7c5c989-8t52p" event={"ID":"8f339b4d-9773-4b73-8828-7cbe31882ff7","Type":"ContainerStarted","Data":"fe38323aa3e661b6285e056850d733fecd56bf524b0c2ac282c17e48d5c8ace5"} Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.062994 4733 generic.go:334] "Generic (PLEG): container finished" podID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerID="f31aa29f94996c0ece09b0fa003c32cb17412c5b29b02e1ac58ae7b31cf6a428" exitCode=0 Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.063048 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a09af527-a663-4dfc-89f0-3e0d3f41c048","Type":"ContainerDied","Data":"f31aa29f94996c0ece09b0fa003c32cb17412c5b29b02e1ac58ae7b31cf6a428"} Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.596889 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.790874 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-combined-ca-bundle\") pod \"a09af527-a663-4dfc-89f0-3e0d3f41c048\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.791405 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-scripts\") pod \"a09af527-a663-4dfc-89f0-3e0d3f41c048\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.791462 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-config-data\") pod \"a09af527-a663-4dfc-89f0-3e0d3f41c048\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.791579 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6qw8\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-kube-api-access-n6qw8\") pod \"a09af527-a663-4dfc-89f0-3e0d3f41c048\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.791665 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-httpd-run\") pod \"a09af527-a663-4dfc-89f0-3e0d3f41c048\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.791686 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-logs\") pod \"a09af527-a663-4dfc-89f0-3e0d3f41c048\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.791783 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-ceph\") pod \"a09af527-a663-4dfc-89f0-3e0d3f41c048\" (UID: \"a09af527-a663-4dfc-89f0-3e0d3f41c048\") " Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.798946 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-scripts" (OuterVolumeSpecName: "scripts") pod "a09af527-a663-4dfc-89f0-3e0d3f41c048" (UID: "a09af527-a663-4dfc-89f0-3e0d3f41c048"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.799636 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-kube-api-access-n6qw8" (OuterVolumeSpecName: "kube-api-access-n6qw8") pod "a09af527-a663-4dfc-89f0-3e0d3f41c048" (UID: "a09af527-a663-4dfc-89f0-3e0d3f41c048"). InnerVolumeSpecName "kube-api-access-n6qw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.894387 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:48 crc kubenswrapper[4733]: I1204 19:22:48.894422 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6qw8\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-kube-api-access-n6qw8\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.079483 4733 generic.go:334] "Generic (PLEG): container finished" podID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerID="1e0372feada67ec46db547f1eb5f49dd406f57ccd058b6aefa078af1a4c96559" exitCode=0 Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.079563 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdf21970-8b52-4c07-81aa-d786baa2bb58","Type":"ContainerDied","Data":"1e0372feada67ec46db547f1eb5f49dd406f57ccd058b6aefa078af1a4c96559"} Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.082205 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a09af527-a663-4dfc-89f0-3e0d3f41c048","Type":"ContainerDied","Data":"e518aee5067eec712fb3c2abf782c9c90583250da14e8d05b867fea3fc7de119"} Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.082254 4733 scope.go:117] "RemoveContainer" containerID="f31aa29f94996c0ece09b0fa003c32cb17412c5b29b02e1ac58ae7b31cf6a428" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.082863 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.543225 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a09af527-a663-4dfc-89f0-3e0d3f41c048" (UID: "a09af527-a663-4dfc-89f0-3e0d3f41c048"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.543258 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-logs" (OuterVolumeSpecName: "logs") pod "a09af527-a663-4dfc-89f0-3e0d3f41c048" (UID: "a09af527-a663-4dfc-89f0-3e0d3f41c048"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.554121 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-ceph" (OuterVolumeSpecName: "ceph") pod "a09af527-a663-4dfc-89f0-3e0d3f41c048" (UID: "a09af527-a663-4dfc-89f0-3e0d3f41c048"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.567709 4733 scope.go:117] "RemoveContainer" containerID="3998038ce1e96df93e35f5935b8e0f2681e565a155bc670d2d3a88053dbec451" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.577038 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a09af527-a663-4dfc-89f0-3e0d3f41c048" (UID: "a09af527-a663-4dfc-89f0-3e0d3f41c048"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.613028 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-config-data" (OuterVolumeSpecName: "config-data") pod "a09af527-a663-4dfc-89f0-3e0d3f41c048" (UID: "a09af527-a663-4dfc-89f0-3e0d3f41c048"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.616006 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.616038 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a09af527-a663-4dfc-89f0-3e0d3f41c048-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.616052 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.616065 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a09af527-a663-4dfc-89f0-3e0d3f41c048-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.616077 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a09af527-a663-4dfc-89f0-3e0d3f41c048-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.742011 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.765968 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.782426 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:22:49 crc kubenswrapper[4733]: E1204 19:22:49.782998 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerName="glance-httpd" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.783025 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerName="glance-httpd" Dec 04 19:22:49 crc kubenswrapper[4733]: E1204 19:22:49.783088 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerName="glance-log" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.783098 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerName="glance-log" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.783340 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerName="glance-log" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.783357 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" containerName="glance-httpd" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.784642 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.787366 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.796544 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.821251 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.821307 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.821348 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.821394 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.821429 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-logs\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.821467 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.821516 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4vtk\" (UniqueName: \"kubernetes.io/projected/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-kube-api-access-r4vtk\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.924108 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4vtk\" (UniqueName: \"kubernetes.io/projected/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-kube-api-access-r4vtk\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.924179 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.924220 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.924278 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.924343 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.924387 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-logs\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.924438 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.925991 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-logs\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.926899 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.930758 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.932091 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.933765 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.944616 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4vtk\" (UniqueName: \"kubernetes.io/projected/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-kube-api-access-r4vtk\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:49 crc kubenswrapper[4733]: I1204 19:22:49.954281 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b95d20bf-4ad1-438a-80e2-34fbac20bb7e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b95d20bf-4ad1-438a-80e2-34fbac20bb7e\") " pod="openstack/glance-default-internal-api-0" Dec 04 19:22:50 crc kubenswrapper[4733]: I1204 19:22:50.132408 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 19:22:50 crc kubenswrapper[4733]: I1204 19:22:50.351282 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a09af527-a663-4dfc-89f0-3e0d3f41c048" path="/var/lib/kubelet/pods/a09af527-a663-4dfc-89f0-3e0d3f41c048/volumes" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.454524 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.522879 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-combined-ca-bundle\") pod \"cdf21970-8b52-4c07-81aa-d786baa2bb58\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.523203 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-ceph\") pod \"cdf21970-8b52-4c07-81aa-d786baa2bb58\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.523259 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-logs\") pod \"cdf21970-8b52-4c07-81aa-d786baa2bb58\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.523355 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvgdj\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-kube-api-access-lvgdj\") pod \"cdf21970-8b52-4c07-81aa-d786baa2bb58\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.523391 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-httpd-run\") pod \"cdf21970-8b52-4c07-81aa-d786baa2bb58\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.523492 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-scripts\") pod \"cdf21970-8b52-4c07-81aa-d786baa2bb58\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.523518 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-config-data\") pod \"cdf21970-8b52-4c07-81aa-d786baa2bb58\" (UID: \"cdf21970-8b52-4c07-81aa-d786baa2bb58\") " Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.527400 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-logs" (OuterVolumeSpecName: "logs") pod "cdf21970-8b52-4c07-81aa-d786baa2bb58" (UID: "cdf21970-8b52-4c07-81aa-d786baa2bb58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.528771 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cdf21970-8b52-4c07-81aa-d786baa2bb58" (UID: "cdf21970-8b52-4c07-81aa-d786baa2bb58"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.535150 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-kube-api-access-lvgdj" (OuterVolumeSpecName: "kube-api-access-lvgdj") pod "cdf21970-8b52-4c07-81aa-d786baa2bb58" (UID: "cdf21970-8b52-4c07-81aa-d786baa2bb58"). InnerVolumeSpecName "kube-api-access-lvgdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.534976 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-ceph" (OuterVolumeSpecName: "ceph") pod "cdf21970-8b52-4c07-81aa-d786baa2bb58" (UID: "cdf21970-8b52-4c07-81aa-d786baa2bb58"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.535572 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-scripts" (OuterVolumeSpecName: "scripts") pod "cdf21970-8b52-4c07-81aa-d786baa2bb58" (UID: "cdf21970-8b52-4c07-81aa-d786baa2bb58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.607547 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdf21970-8b52-4c07-81aa-d786baa2bb58" (UID: "cdf21970-8b52-4c07-81aa-d786baa2bb58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.628921 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.628951 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvgdj\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-kube-api-access-lvgdj\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.628960 4733 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdf21970-8b52-4c07-81aa-d786baa2bb58-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.628969 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.628977 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.628984 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cdf21970-8b52-4c07-81aa-d786baa2bb58-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.652547 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-config-data" (OuterVolumeSpecName: "config-data") pod "cdf21970-8b52-4c07-81aa-d786baa2bb58" (UID: "cdf21970-8b52-4c07-81aa-d786baa2bb58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.730895 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf21970-8b52-4c07-81aa-d786baa2bb58-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:22:54 crc kubenswrapper[4733]: I1204 19:22:54.891561 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 19:22:54 crc kubenswrapper[4733]: W1204 19:22:54.923920 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb95d20bf_4ad1_438a_80e2_34fbac20bb7e.slice/crio-e264948d704cb036f305d9ca0f935e369312a6c4d20380212659173923031995 WatchSource:0}: Error finding container e264948d704cb036f305d9ca0f935e369312a6c4d20380212659173923031995: Status 404 returned error can't find the container with id e264948d704cb036f305d9ca0f935e369312a6c4d20380212659173923031995 Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.175640 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d7c5c989-8t52p" event={"ID":"8f339b4d-9773-4b73-8828-7cbe31882ff7","Type":"ContainerStarted","Data":"63f0449d7ee89d9ebb20cd215985fea362dafbff08e29ebb67213c03e5e2846c"} Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.175689 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d7c5c989-8t52p" event={"ID":"8f339b4d-9773-4b73-8828-7cbe31882ff7","Type":"ContainerStarted","Data":"0b69c74c8f7594a7cc51e4de12f8e94fcfe817a7afa5f0aee1f4be5669d9bb32"} Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.177277 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b95d20bf-4ad1-438a-80e2-34fbac20bb7e","Type":"ContainerStarted","Data":"e264948d704cb036f305d9ca0f935e369312a6c4d20380212659173923031995"} Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.179868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-989dbf8fc-q2crf" event={"ID":"648988d6-2223-4c06-b39f-c899a2778082","Type":"ContainerStarted","Data":"a21d7cd81789636e821ed3a2ace578ada3f0add82ba799370a593fc028d5a66a"} Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.179917 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-989dbf8fc-q2crf" podUID="648988d6-2223-4c06-b39f-c899a2778082" containerName="horizon-log" containerID="cri-o://a2983d8c75c7e3fcad745d5a5baf4a79c8dd1311551c099400fbd249fc5dea54" gracePeriod=30 Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.179956 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-989dbf8fc-q2crf" event={"ID":"648988d6-2223-4c06-b39f-c899a2778082","Type":"ContainerStarted","Data":"a2983d8c75c7e3fcad745d5a5baf4a79c8dd1311551c099400fbd249fc5dea54"} Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.179937 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-989dbf8fc-q2crf" podUID="648988d6-2223-4c06-b39f-c899a2778082" containerName="horizon" containerID="cri-o://a21d7cd81789636e821ed3a2ace578ada3f0add82ba799370a593fc028d5a66a" gracePeriod=30 Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.189366 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdf21970-8b52-4c07-81aa-d786baa2bb58","Type":"ContainerDied","Data":"916df4214a2bf157cb3b95fcbd454ee09d463de0d5706a212eb8cf6a6ed01406"} Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.189434 4733 scope.go:117] "RemoveContainer" containerID="1e0372feada67ec46db547f1eb5f49dd406f57ccd058b6aefa078af1a4c96559" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.189615 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.197324 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d8bc57f7-p95j8" event={"ID":"66d42413-6ae0-4845-b25c-4d19058854f5","Type":"ContainerStarted","Data":"f3411c5af1bcd50e9844b7232f6d9a0fbde9b34377784d4c9df3831afffa4e99"} Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.197413 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d8bc57f7-p95j8" event={"ID":"66d42413-6ae0-4845-b25c-4d19058854f5","Type":"ContainerStarted","Data":"6e8e04ac655bc1e690d341e87f55fd2542bdb0cb52355f57a6d5f6a875cca5dd"} Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.214829 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.220200 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66d7c5c989-8t52p" podStartSLOduration=2.134748716 podStartE2EDuration="10.220179847s" podCreationTimestamp="2025-12-04 19:22:45 +0000 UTC" firstStartedPulling="2025-12-04 19:22:46.357527538 +0000 UTC m=+6228.312888584" lastFinishedPulling="2025-12-04 19:22:54.442958669 +0000 UTC m=+6236.398319715" observedRunningTime="2025-12-04 19:22:55.202590791 +0000 UTC m=+6237.157951827" watchObservedRunningTime="2025-12-04 19:22:55.220179847 +0000 UTC m=+6237.175540893" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.235274 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-9d8bc57f7-p95j8" podStartSLOduration=2.645165477 podStartE2EDuration="11.235252815s" podCreationTimestamp="2025-12-04 19:22:44 +0000 UTC" firstStartedPulling="2025-12-04 19:22:45.811695716 +0000 UTC m=+6227.767056752" lastFinishedPulling="2025-12-04 19:22:54.401783044 +0000 UTC m=+6236.357144090" observedRunningTime="2025-12-04 19:22:55.226879588 +0000 UTC m=+6237.182240624" watchObservedRunningTime="2025-12-04 19:22:55.235252815 +0000 UTC m=+6237.190613861" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.239784 4733 scope.go:117] "RemoveContainer" containerID="de21d2bb957d01d50a27342f3b90c33caf498e96ba91ca2639f26e7c0dccc6c4" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.268184 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-989dbf8fc-q2crf" podStartSLOduration=2.7370697760000002 podStartE2EDuration="11.268157936s" podCreationTimestamp="2025-12-04 19:22:44 +0000 UTC" firstStartedPulling="2025-12-04 19:22:45.874372294 +0000 UTC m=+6227.829733340" lastFinishedPulling="2025-12-04 19:22:54.405460444 +0000 UTC m=+6236.360821500" observedRunningTime="2025-12-04 19:22:55.245149844 +0000 UTC m=+6237.200510890" watchObservedRunningTime="2025-12-04 19:22:55.268157936 +0000 UTC m=+6237.223518982" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.305864 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.315104 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.327107 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:22:55 crc kubenswrapper[4733]: E1204 19:22:55.327566 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerName="glance-httpd" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.327588 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerName="glance-httpd" Dec 04 19:22:55 crc kubenswrapper[4733]: E1204 19:22:55.327614 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerName="glance-log" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.327622 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerName="glance-log" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.327922 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerName="glance-log" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.327948 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" containerName="glance-httpd" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.329130 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.332901 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.345908 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-scripts\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.346031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvd7z\" (UniqueName: \"kubernetes.io/projected/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-kube-api-access-xvd7z\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.346130 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-ceph\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.346156 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-config-data\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.346186 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.346207 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.346378 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-logs\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.363303 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.449039 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-logs\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.449985 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-scripts\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.450221 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvd7z\" (UniqueName: \"kubernetes.io/projected/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-kube-api-access-xvd7z\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.450424 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-ceph\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.450552 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-config-data\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.450716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.451017 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.451199 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.449527 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-logs\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.464819 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-ceph\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.471428 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-config-data\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.472434 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-scripts\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.477980 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.478583 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvd7z\" (UniqueName: \"kubernetes.io/projected/9455ff4a-328e-4b58-bd3e-a3e7b11c7737-kube-api-access-xvd7z\") pod \"glance-default-external-api-0\" (UID: \"9455ff4a-328e-4b58-bd3e-a3e7b11c7737\") " pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.684400 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.966870 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:55 crc kubenswrapper[4733]: I1204 19:22:55.967792 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:22:56 crc kubenswrapper[4733]: I1204 19:22:56.208503 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b95d20bf-4ad1-438a-80e2-34fbac20bb7e","Type":"ContainerStarted","Data":"ca33ddb1f685e4360244cf4c49b04cb9aef44d7a879ed371305b519e07bf52f0"} Dec 04 19:22:56 crc kubenswrapper[4733]: I1204 19:22:56.348590 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf21970-8b52-4c07-81aa-d786baa2bb58" path="/var/lib/kubelet/pods/cdf21970-8b52-4c07-81aa-d786baa2bb58/volumes" Dec 04 19:22:56 crc kubenswrapper[4733]: I1204 19:22:56.392696 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 19:22:57 crc kubenswrapper[4733]: I1204 19:22:57.232359 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b95d20bf-4ad1-438a-80e2-34fbac20bb7e","Type":"ContainerStarted","Data":"f479b66e038464c3cdc4da1dfff48ec4bae2aeac30191e6a45d731cdcfa4c6eb"} Dec 04 19:22:57 crc kubenswrapper[4733]: I1204 19:22:57.240470 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9455ff4a-328e-4b58-bd3e-a3e7b11c7737","Type":"ContainerStarted","Data":"97ea5d1b07a4be76600ea1f519b7624f23b8d20158c69a6ea63fbd18a4d5140e"} Dec 04 19:22:57 crc kubenswrapper[4733]: I1204 19:22:57.240509 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9455ff4a-328e-4b58-bd3e-a3e7b11c7737","Type":"ContainerStarted","Data":"bece6fb255e598ccde65dee91600a991544f28684d20e3c9a16b44bbf72d8a29"} Dec 04 19:22:57 crc kubenswrapper[4733]: I1204 19:22:57.257696 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.257676654 podStartE2EDuration="8.257676654s" podCreationTimestamp="2025-12-04 19:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:22:57.252879744 +0000 UTC m=+6239.208240790" watchObservedRunningTime="2025-12-04 19:22:57.257676654 +0000 UTC m=+6239.213037700" Dec 04 19:22:58 crc kubenswrapper[4733]: I1204 19:22:58.253630 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9455ff4a-328e-4b58-bd3e-a3e7b11c7737","Type":"ContainerStarted","Data":"829bd5e25f39e5db0673b5bb18a70a7552b8d74a25f163ef6d665fd789156be8"} Dec 04 19:22:58 crc kubenswrapper[4733]: I1204 19:22:58.295289 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.295266594 podStartE2EDuration="3.295266594s" podCreationTimestamp="2025-12-04 19:22:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:22:58.280188375 +0000 UTC m=+6240.235549431" watchObservedRunningTime="2025-12-04 19:22:58.295266594 +0000 UTC m=+6240.250627640" Dec 04 19:23:00 crc kubenswrapper[4733]: I1204 19:23:00.134108 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 19:23:00 crc kubenswrapper[4733]: I1204 19:23:00.136395 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 19:23:00 crc kubenswrapper[4733]: I1204 19:23:00.179189 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 19:23:00 crc kubenswrapper[4733]: I1204 19:23:00.183780 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 19:23:00 crc kubenswrapper[4733]: I1204 19:23:00.273732 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 19:23:00 crc kubenswrapper[4733]: I1204 19:23:00.274142 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 19:23:01 crc kubenswrapper[4733]: I1204 19:23:01.083507 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-53c8-account-create-update-hgqsk"] Dec 04 19:23:01 crc kubenswrapper[4733]: I1204 19:23:01.092885 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-8t6f2"] Dec 04 19:23:01 crc kubenswrapper[4733]: I1204 19:23:01.104864 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-8t6f2"] Dec 04 19:23:01 crc kubenswrapper[4733]: I1204 19:23:01.109510 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-53c8-account-create-update-hgqsk"] Dec 04 19:23:02 crc kubenswrapper[4733]: I1204 19:23:02.349766 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15fb0119-f4e2-4e49-a652-5306d9b0a622" path="/var/lib/kubelet/pods/15fb0119-f4e2-4e49-a652-5306d9b0a622/volumes" Dec 04 19:23:02 crc kubenswrapper[4733]: I1204 19:23:02.352454 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31826b3c-ded5-4174-9796-3b8d4aa4abb4" path="/var/lib/kubelet/pods/31826b3c-ded5-4174-9796-3b8d4aa4abb4/volumes" Dec 04 19:23:02 crc kubenswrapper[4733]: I1204 19:23:02.862507 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 19:23:03 crc kubenswrapper[4733]: I1204 19:23:03.307587 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 19:23:05 crc kubenswrapper[4733]: I1204 19:23:05.057495 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:23:05 crc kubenswrapper[4733]: I1204 19:23:05.059122 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:23:05 crc kubenswrapper[4733]: I1204 19:23:05.064120 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-9d8bc57f7-p95j8" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 04 19:23:05 crc kubenswrapper[4733]: I1204 19:23:05.685111 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 19:23:05 crc kubenswrapper[4733]: I1204 19:23:05.685583 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 19:23:05 crc kubenswrapper[4733]: I1204 19:23:05.719675 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 19:23:05 crc kubenswrapper[4733]: I1204 19:23:05.748692 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 19:23:05 crc kubenswrapper[4733]: I1204 19:23:05.896544 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66d7c5c989-8t52p" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.117:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8080: connect: connection refused" Dec 04 19:23:06 crc kubenswrapper[4733]: I1204 19:23:06.347037 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 19:23:06 crc kubenswrapper[4733]: I1204 19:23:06.347067 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 19:23:08 crc kubenswrapper[4733]: I1204 19:23:08.307324 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 19:23:08 crc kubenswrapper[4733]: I1204 19:23:08.309826 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 19:23:09 crc kubenswrapper[4733]: I1204 19:23:09.029419 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2vgfw"] Dec 04 19:23:09 crc kubenswrapper[4733]: I1204 19:23:09.038067 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2vgfw"] Dec 04 19:23:10 crc kubenswrapper[4733]: I1204 19:23:10.347579 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaaf0141-6b6c-4e56-8fc1-c7a875c3be77" path="/var/lib/kubelet/pods/eaaf0141-6b6c-4e56-8fc1-c7a875c3be77/volumes" Dec 04 19:23:15 crc kubenswrapper[4733]: I1204 19:23:15.362119 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:23:15 crc kubenswrapper[4733]: I1204 19:23:15.362585 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:23:16 crc kubenswrapper[4733]: I1204 19:23:16.944363 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:23:17 crc kubenswrapper[4733]: I1204 19:23:17.606621 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:23:18 crc kubenswrapper[4733]: I1204 19:23:18.674892 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:23:19 crc kubenswrapper[4733]: I1204 19:23:19.313285 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:23:19 crc kubenswrapper[4733]: I1204 19:23:19.365881 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9d8bc57f7-p95j8"] Dec 04 19:23:19 crc kubenswrapper[4733]: I1204 19:23:19.498659 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-9d8bc57f7-p95j8" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon-log" containerID="cri-o://6e8e04ac655bc1e690d341e87f55fd2542bdb0cb52355f57a6d5f6a875cca5dd" gracePeriod=30 Dec 04 19:23:19 crc kubenswrapper[4733]: I1204 19:23:19.498823 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-9d8bc57f7-p95j8" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon" containerID="cri-o://f3411c5af1bcd50e9844b7232f6d9a0fbde9b34377784d4c9df3831afffa4e99" gracePeriod=30 Dec 04 19:23:23 crc kubenswrapper[4733]: I1204 19:23:23.294286 4733 scope.go:117] "RemoveContainer" containerID="084acd82ccfad183284260343c3ddeeea95837054db447011fdf86650f727c66" Dec 04 19:23:23 crc kubenswrapper[4733]: I1204 19:23:23.331687 4733 scope.go:117] "RemoveContainer" containerID="950263320cd4ce323201835602127727776b8fdec7eb2e49ce2e935611928b1c" Dec 04 19:23:23 crc kubenswrapper[4733]: I1204 19:23:23.378572 4733 scope.go:117] "RemoveContainer" containerID="9b1ada6de779a7c2d39facc56c0642fe857486bc6e61159cd487d610f5ed8ed3" Dec 04 19:23:23 crc kubenswrapper[4733]: I1204 19:23:23.414156 4733 scope.go:117] "RemoveContainer" containerID="9fb8361d9617c8ab101d44217631084e02369a3904f663990070e5df1e5f69b0" Dec 04 19:23:23 crc kubenswrapper[4733]: I1204 19:23:23.458665 4733 scope.go:117] "RemoveContainer" containerID="98a4162d8e1bd916f1f76ec054ae8304f012c17dbc2741a3a956ef56f37882a8" Dec 04 19:23:23 crc kubenswrapper[4733]: I1204 19:23:23.527764 4733 scope.go:117] "RemoveContainer" containerID="3669ce59afc64860ea6608339208ab3b75b9c7fcb25f0e1d5c3c83f760b4eea9" Dec 04 19:23:23 crc kubenswrapper[4733]: I1204 19:23:23.545493 4733 generic.go:334] "Generic (PLEG): container finished" podID="66d42413-6ae0-4845-b25c-4d19058854f5" containerID="f3411c5af1bcd50e9844b7232f6d9a0fbde9b34377784d4c9df3831afffa4e99" exitCode=0 Dec 04 19:23:23 crc kubenswrapper[4733]: I1204 19:23:23.545529 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d8bc57f7-p95j8" event={"ID":"66d42413-6ae0-4845-b25c-4d19058854f5","Type":"ContainerDied","Data":"f3411c5af1bcd50e9844b7232f6d9a0fbde9b34377784d4c9df3831afffa4e99"} Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.058516 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-9d8bc57f7-p95j8" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.569325 4733 generic.go:334] "Generic (PLEG): container finished" podID="648988d6-2223-4c06-b39f-c899a2778082" containerID="a21d7cd81789636e821ed3a2ace578ada3f0add82ba799370a593fc028d5a66a" exitCode=137 Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.569631 4733 generic.go:334] "Generic (PLEG): container finished" podID="648988d6-2223-4c06-b39f-c899a2778082" containerID="a2983d8c75c7e3fcad745d5a5baf4a79c8dd1311551c099400fbd249fc5dea54" exitCode=137 Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.569412 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-989dbf8fc-q2crf" event={"ID":"648988d6-2223-4c06-b39f-c899a2778082","Type":"ContainerDied","Data":"a21d7cd81789636e821ed3a2ace578ada3f0add82ba799370a593fc028d5a66a"} Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.569677 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-989dbf8fc-q2crf" event={"ID":"648988d6-2223-4c06-b39f-c899a2778082","Type":"ContainerDied","Data":"a2983d8c75c7e3fcad745d5a5baf4a79c8dd1311551c099400fbd249fc5dea54"} Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.569695 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-989dbf8fc-q2crf" event={"ID":"648988d6-2223-4c06-b39f-c899a2778082","Type":"ContainerDied","Data":"0d51de6cb0e2e9c8f29954909c3f23ded6d96dd3c7133ec9fdcdce9096583ac6"} Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.569709 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d51de6cb0e2e9c8f29954909c3f23ded6d96dd3c7133ec9fdcdce9096583ac6" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.636100 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.702990 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/648988d6-2223-4c06-b39f-c899a2778082-logs\") pod \"648988d6-2223-4c06-b39f-c899a2778082\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.703086 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmxw4\" (UniqueName: \"kubernetes.io/projected/648988d6-2223-4c06-b39f-c899a2778082-kube-api-access-zmxw4\") pod \"648988d6-2223-4c06-b39f-c899a2778082\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.703272 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/648988d6-2223-4c06-b39f-c899a2778082-horizon-secret-key\") pod \"648988d6-2223-4c06-b39f-c899a2778082\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.703340 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-scripts\") pod \"648988d6-2223-4c06-b39f-c899a2778082\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.703370 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-config-data\") pod \"648988d6-2223-4c06-b39f-c899a2778082\" (UID: \"648988d6-2223-4c06-b39f-c899a2778082\") " Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.704475 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/648988d6-2223-4c06-b39f-c899a2778082-logs" (OuterVolumeSpecName: "logs") pod "648988d6-2223-4c06-b39f-c899a2778082" (UID: "648988d6-2223-4c06-b39f-c899a2778082"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.705011 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/648988d6-2223-4c06-b39f-c899a2778082-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.709403 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/648988d6-2223-4c06-b39f-c899a2778082-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "648988d6-2223-4c06-b39f-c899a2778082" (UID: "648988d6-2223-4c06-b39f-c899a2778082"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.723119 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/648988d6-2223-4c06-b39f-c899a2778082-kube-api-access-zmxw4" (OuterVolumeSpecName: "kube-api-access-zmxw4") pod "648988d6-2223-4c06-b39f-c899a2778082" (UID: "648988d6-2223-4c06-b39f-c899a2778082"). InnerVolumeSpecName "kube-api-access-zmxw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.737262 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-config-data" (OuterVolumeSpecName: "config-data") pod "648988d6-2223-4c06-b39f-c899a2778082" (UID: "648988d6-2223-4c06-b39f-c899a2778082"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.740297 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-scripts" (OuterVolumeSpecName: "scripts") pod "648988d6-2223-4c06-b39f-c899a2778082" (UID: "648988d6-2223-4c06-b39f-c899a2778082"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.806962 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.806988 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/648988d6-2223-4c06-b39f-c899a2778082-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.807000 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmxw4\" (UniqueName: \"kubernetes.io/projected/648988d6-2223-4c06-b39f-c899a2778082-kube-api-access-zmxw4\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:25 crc kubenswrapper[4733]: I1204 19:23:25.807009 4733 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/648988d6-2223-4c06-b39f-c899a2778082-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:26 crc kubenswrapper[4733]: I1204 19:23:26.581933 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-989dbf8fc-q2crf" Dec 04 19:23:26 crc kubenswrapper[4733]: I1204 19:23:26.636942 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-989dbf8fc-q2crf"] Dec 04 19:23:26 crc kubenswrapper[4733]: I1204 19:23:26.642843 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-989dbf8fc-q2crf"] Dec 04 19:23:28 crc kubenswrapper[4733]: I1204 19:23:28.353292 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="648988d6-2223-4c06-b39f-c899a2778082" path="/var/lib/kubelet/pods/648988d6-2223-4c06-b39f-c899a2778082/volumes" Dec 04 19:23:35 crc kubenswrapper[4733]: I1204 19:23:35.058480 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-9d8bc57f7-p95j8" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.058536 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-9d8bc57f7-p95j8" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.059384 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.362216 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.362292 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.362350 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.363367 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"353758af15d16464418cca9546b4d8b6ebe8b41cef4796d49135ac10e1a77f6b"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.363471 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://353758af15d16464418cca9546b4d8b6ebe8b41cef4796d49135ac10e1a77f6b" gracePeriod=600 Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.818781 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="353758af15d16464418cca9546b4d8b6ebe8b41cef4796d49135ac10e1a77f6b" exitCode=0 Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.818882 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"353758af15d16464418cca9546b4d8b6ebe8b41cef4796d49135ac10e1a77f6b"} Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.819200 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08"} Dec 04 19:23:45 crc kubenswrapper[4733]: I1204 19:23:45.819224 4733 scope.go:117] "RemoveContainer" containerID="b8fe8125edb495222e80e766efe4bd54625a542a9e8cde44edf18ed6ada43fc0" Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.869749 4733 generic.go:334] "Generic (PLEG): container finished" podID="66d42413-6ae0-4845-b25c-4d19058854f5" containerID="6e8e04ac655bc1e690d341e87f55fd2542bdb0cb52355f57a6d5f6a875cca5dd" exitCode=137 Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.869875 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d8bc57f7-p95j8" event={"ID":"66d42413-6ae0-4845-b25c-4d19058854f5","Type":"ContainerDied","Data":"6e8e04ac655bc1e690d341e87f55fd2542bdb0cb52355f57a6d5f6a875cca5dd"} Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.870354 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d8bc57f7-p95j8" event={"ID":"66d42413-6ae0-4845-b25c-4d19058854f5","Type":"ContainerDied","Data":"c09e972ecce60889651da09566a2b851d1ade38412fb02ab7d89f5c106af1e3b"} Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.870374 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c09e972ecce60889651da09566a2b851d1ade38412fb02ab7d89f5c106af1e3b" Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.932413 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.960790 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66d42413-6ae0-4845-b25c-4d19058854f5-horizon-secret-key\") pod \"66d42413-6ae0-4845-b25c-4d19058854f5\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.960912 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqlnz\" (UniqueName: \"kubernetes.io/projected/66d42413-6ae0-4845-b25c-4d19058854f5-kube-api-access-sqlnz\") pod \"66d42413-6ae0-4845-b25c-4d19058854f5\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.960954 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-scripts\") pod \"66d42413-6ae0-4845-b25c-4d19058854f5\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.961138 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d42413-6ae0-4845-b25c-4d19058854f5-logs\") pod \"66d42413-6ae0-4845-b25c-4d19058854f5\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.961171 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-config-data\") pod \"66d42413-6ae0-4845-b25c-4d19058854f5\" (UID: \"66d42413-6ae0-4845-b25c-4d19058854f5\") " Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.962224 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d42413-6ae0-4845-b25c-4d19058854f5-logs" (OuterVolumeSpecName: "logs") pod "66d42413-6ae0-4845-b25c-4d19058854f5" (UID: "66d42413-6ae0-4845-b25c-4d19058854f5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.967505 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d42413-6ae0-4845-b25c-4d19058854f5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "66d42413-6ae0-4845-b25c-4d19058854f5" (UID: "66d42413-6ae0-4845-b25c-4d19058854f5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.967658 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d42413-6ae0-4845-b25c-4d19058854f5-kube-api-access-sqlnz" (OuterVolumeSpecName: "kube-api-access-sqlnz") pod "66d42413-6ae0-4845-b25c-4d19058854f5" (UID: "66d42413-6ae0-4845-b25c-4d19058854f5"). InnerVolumeSpecName "kube-api-access-sqlnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.996037 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-scripts" (OuterVolumeSpecName: "scripts") pod "66d42413-6ae0-4845-b25c-4d19058854f5" (UID: "66d42413-6ae0-4845-b25c-4d19058854f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:23:49 crc kubenswrapper[4733]: I1204 19:23:49.997681 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-config-data" (OuterVolumeSpecName: "config-data") pod "66d42413-6ae0-4845-b25c-4d19058854f5" (UID: "66d42413-6ae0-4845-b25c-4d19058854f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:23:50 crc kubenswrapper[4733]: I1204 19:23:50.064705 4733 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66d42413-6ae0-4845-b25c-4d19058854f5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:50 crc kubenswrapper[4733]: I1204 19:23:50.064735 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqlnz\" (UniqueName: \"kubernetes.io/projected/66d42413-6ae0-4845-b25c-4d19058854f5-kube-api-access-sqlnz\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:50 crc kubenswrapper[4733]: I1204 19:23:50.064747 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:50 crc kubenswrapper[4733]: I1204 19:23:50.064754 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d42413-6ae0-4845-b25c-4d19058854f5-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:50 crc kubenswrapper[4733]: I1204 19:23:50.064763 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66d42413-6ae0-4845-b25c-4d19058854f5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:23:50 crc kubenswrapper[4733]: I1204 19:23:50.887991 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9d8bc57f7-p95j8" Dec 04 19:23:50 crc kubenswrapper[4733]: I1204 19:23:50.947392 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9d8bc57f7-p95j8"] Dec 04 19:23:50 crc kubenswrapper[4733]: I1204 19:23:50.960413 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-9d8bc57f7-p95j8"] Dec 04 19:23:51 crc kubenswrapper[4733]: I1204 19:23:51.074305 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b55f-account-create-update-5mslq"] Dec 04 19:23:51 crc kubenswrapper[4733]: I1204 19:23:51.086606 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-zf7qx"] Dec 04 19:23:51 crc kubenswrapper[4733]: I1204 19:23:51.097968 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b55f-account-create-update-5mslq"] Dec 04 19:23:51 crc kubenswrapper[4733]: I1204 19:23:51.107705 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-zf7qx"] Dec 04 19:23:52 crc kubenswrapper[4733]: I1204 19:23:52.363708 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" path="/var/lib/kubelet/pods/66d42413-6ae0-4845-b25c-4d19058854f5/volumes" Dec 04 19:23:52 crc kubenswrapper[4733]: I1204 19:23:52.365140 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd069278-c6ab-4341-a9ed-1b93d4aa6280" path="/var/lib/kubelet/pods/dd069278-c6ab-4341-a9ed-1b93d4aa6280/volumes" Dec 04 19:23:52 crc kubenswrapper[4733]: I1204 19:23:52.365997 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e986a553-d439-4add-90ea-a3d15d378c29" path="/var/lib/kubelet/pods/e986a553-d439-4add-90ea-a3d15d378c29/volumes" Dec 04 19:24:00 crc kubenswrapper[4733]: I1204 19:24:00.040914 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-mgml8"] Dec 04 19:24:00 crc kubenswrapper[4733]: I1204 19:24:00.056189 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-mgml8"] Dec 04 19:24:00 crc kubenswrapper[4733]: I1204 19:24:00.346949 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d88cdb3-e89d-4743-a918-4485083caaf5" path="/var/lib/kubelet/pods/0d88cdb3-e89d-4743-a918-4485083caaf5/volumes" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.329848 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5884fc67b7-6bjb4"] Dec 04 19:24:02 crc kubenswrapper[4733]: E1204 19:24:02.330620 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon-log" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.330635 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon-log" Dec 04 19:24:02 crc kubenswrapper[4733]: E1204 19:24:02.330654 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="648988d6-2223-4c06-b39f-c899a2778082" containerName="horizon-log" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.330661 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="648988d6-2223-4c06-b39f-c899a2778082" containerName="horizon-log" Dec 04 19:24:02 crc kubenswrapper[4733]: E1204 19:24:02.330677 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="648988d6-2223-4c06-b39f-c899a2778082" containerName="horizon" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.330684 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="648988d6-2223-4c06-b39f-c899a2778082" containerName="horizon" Dec 04 19:24:02 crc kubenswrapper[4733]: E1204 19:24:02.330703 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.330710 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.330967 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="648988d6-2223-4c06-b39f-c899a2778082" containerName="horizon-log" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.330996 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon-log" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.331019 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="648988d6-2223-4c06-b39f-c899a2778082" containerName="horizon" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.331038 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d42413-6ae0-4845-b25c-4d19058854f5" containerName="horizon" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.332043 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.386304 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5884fc67b7-6bjb4"] Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.446981 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-logs\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.447122 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-scripts\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.447219 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h54gr\" (UniqueName: \"kubernetes.io/projected/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-kube-api-access-h54gr\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.447304 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-config-data\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.447332 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-horizon-secret-key\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.550117 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-logs\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.550299 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-scripts\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.550412 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h54gr\" (UniqueName: \"kubernetes.io/projected/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-kube-api-access-h54gr\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.550546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-config-data\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.550590 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-horizon-secret-key\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.560536 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-logs\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.560927 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-scripts\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.561945 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-config-data\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.603505 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-horizon-secret-key\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.615377 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h54gr\" (UniqueName: \"kubernetes.io/projected/21cdbfda-59ac-4b40-ab39-8efde3ba9de7-kube-api-access-h54gr\") pod \"horizon-5884fc67b7-6bjb4\" (UID: \"21cdbfda-59ac-4b40-ab39-8efde3ba9de7\") " pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:02 crc kubenswrapper[4733]: I1204 19:24:02.692234 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.202578 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5884fc67b7-6bjb4"] Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.628955 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-ft7hm"] Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.630477 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.647394 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-ft7hm"] Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.671836 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7dab29-a923-4ace-811a-cad3279c16ed-operator-scripts\") pod \"heat-db-create-ft7hm\" (UID: \"ed7dab29-a923-4ace-811a-cad3279c16ed\") " pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.671888 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b8j4\" (UniqueName: \"kubernetes.io/projected/ed7dab29-a923-4ace-811a-cad3279c16ed-kube-api-access-4b8j4\") pod \"heat-db-create-ft7hm\" (UID: \"ed7dab29-a923-4ace-811a-cad3279c16ed\") " pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.729349 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-2e40-account-create-update-jtgj4"] Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.730587 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.732305 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.744160 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-2e40-account-create-update-jtgj4"] Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.773711 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k45b4\" (UniqueName: \"kubernetes.io/projected/ef49d872-16fc-4e04-b146-5a7c9b260111-kube-api-access-k45b4\") pod \"heat-2e40-account-create-update-jtgj4\" (UID: \"ef49d872-16fc-4e04-b146-5a7c9b260111\") " pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.773762 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef49d872-16fc-4e04-b146-5a7c9b260111-operator-scripts\") pod \"heat-2e40-account-create-update-jtgj4\" (UID: \"ef49d872-16fc-4e04-b146-5a7c9b260111\") " pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.773986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7dab29-a923-4ace-811a-cad3279c16ed-operator-scripts\") pod \"heat-db-create-ft7hm\" (UID: \"ed7dab29-a923-4ace-811a-cad3279c16ed\") " pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.774019 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b8j4\" (UniqueName: \"kubernetes.io/projected/ed7dab29-a923-4ace-811a-cad3279c16ed-kube-api-access-4b8j4\") pod \"heat-db-create-ft7hm\" (UID: \"ed7dab29-a923-4ace-811a-cad3279c16ed\") " pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.774727 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7dab29-a923-4ace-811a-cad3279c16ed-operator-scripts\") pod \"heat-db-create-ft7hm\" (UID: \"ed7dab29-a923-4ace-811a-cad3279c16ed\") " pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.792613 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b8j4\" (UniqueName: \"kubernetes.io/projected/ed7dab29-a923-4ace-811a-cad3279c16ed-kube-api-access-4b8j4\") pod \"heat-db-create-ft7hm\" (UID: \"ed7dab29-a923-4ace-811a-cad3279c16ed\") " pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.875964 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k45b4\" (UniqueName: \"kubernetes.io/projected/ef49d872-16fc-4e04-b146-5a7c9b260111-kube-api-access-k45b4\") pod \"heat-2e40-account-create-update-jtgj4\" (UID: \"ef49d872-16fc-4e04-b146-5a7c9b260111\") " pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.876019 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef49d872-16fc-4e04-b146-5a7c9b260111-operator-scripts\") pod \"heat-2e40-account-create-update-jtgj4\" (UID: \"ef49d872-16fc-4e04-b146-5a7c9b260111\") " pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.876696 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef49d872-16fc-4e04-b146-5a7c9b260111-operator-scripts\") pod \"heat-2e40-account-create-update-jtgj4\" (UID: \"ef49d872-16fc-4e04-b146-5a7c9b260111\") " pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.892411 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k45b4\" (UniqueName: \"kubernetes.io/projected/ef49d872-16fc-4e04-b146-5a7c9b260111-kube-api-access-k45b4\") pod \"heat-2e40-account-create-update-jtgj4\" (UID: \"ef49d872-16fc-4e04-b146-5a7c9b260111\") " pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:03 crc kubenswrapper[4733]: I1204 19:24:03.966546 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:04 crc kubenswrapper[4733]: I1204 19:24:04.021030 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5884fc67b7-6bjb4" event={"ID":"21cdbfda-59ac-4b40-ab39-8efde3ba9de7","Type":"ContainerStarted","Data":"6911cf2db7964023b0547cd9f367b7d21ceeb357e2e05ad53791636ff00f7e9a"} Dec 04 19:24:04 crc kubenswrapper[4733]: I1204 19:24:04.021072 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5884fc67b7-6bjb4" event={"ID":"21cdbfda-59ac-4b40-ab39-8efde3ba9de7","Type":"ContainerStarted","Data":"3c16a245a9e980310f7514524376b5cc64691de161a2abc53a133d88c43b2c5d"} Dec 04 19:24:04 crc kubenswrapper[4733]: I1204 19:24:04.021082 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5884fc67b7-6bjb4" event={"ID":"21cdbfda-59ac-4b40-ab39-8efde3ba9de7","Type":"ContainerStarted","Data":"598f7aa61cb683fe4b8b250e130397e50c5e7a641d1ec78cb08bbb83f760cae7"} Dec 04 19:24:04 crc kubenswrapper[4733]: I1204 19:24:04.048401 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:04 crc kubenswrapper[4733]: I1204 19:24:04.054985 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5884fc67b7-6bjb4" podStartSLOduration=2.054966309 podStartE2EDuration="2.054966309s" podCreationTimestamp="2025-12-04 19:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:24:04.042942433 +0000 UTC m=+6305.998303479" watchObservedRunningTime="2025-12-04 19:24:04.054966309 +0000 UTC m=+6306.010327355" Dec 04 19:24:04 crc kubenswrapper[4733]: I1204 19:24:04.509304 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-ft7hm"] Dec 04 19:24:04 crc kubenswrapper[4733]: W1204 19:24:04.511888 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded7dab29_a923_4ace_811a_cad3279c16ed.slice/crio-09fa863decd8b43d48b1bb7d548f5e0b2fcece81444c575d7aff50fa57eedc1a WatchSource:0}: Error finding container 09fa863decd8b43d48b1bb7d548f5e0b2fcece81444c575d7aff50fa57eedc1a: Status 404 returned error can't find the container with id 09fa863decd8b43d48b1bb7d548f5e0b2fcece81444c575d7aff50fa57eedc1a Dec 04 19:24:04 crc kubenswrapper[4733]: I1204 19:24:04.599511 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-2e40-account-create-update-jtgj4"] Dec 04 19:24:04 crc kubenswrapper[4733]: W1204 19:24:04.612638 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef49d872_16fc_4e04_b146_5a7c9b260111.slice/crio-b194b9c91febd25d7e6292f8e7d4e8d8fcb4cb1c4b82e1802cb409198463812d WatchSource:0}: Error finding container b194b9c91febd25d7e6292f8e7d4e8d8fcb4cb1c4b82e1802cb409198463812d: Status 404 returned error can't find the container with id b194b9c91febd25d7e6292f8e7d4e8d8fcb4cb1c4b82e1802cb409198463812d Dec 04 19:24:05 crc kubenswrapper[4733]: I1204 19:24:05.030980 4733 generic.go:334] "Generic (PLEG): container finished" podID="ed7dab29-a923-4ace-811a-cad3279c16ed" containerID="79d76ad7051f1d82b322e12e457c1bfe302b4dd31065fc2ef32982d15f5ed37e" exitCode=0 Dec 04 19:24:05 crc kubenswrapper[4733]: I1204 19:24:05.031054 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ft7hm" event={"ID":"ed7dab29-a923-4ace-811a-cad3279c16ed","Type":"ContainerDied","Data":"79d76ad7051f1d82b322e12e457c1bfe302b4dd31065fc2ef32982d15f5ed37e"} Dec 04 19:24:05 crc kubenswrapper[4733]: I1204 19:24:05.031389 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ft7hm" event={"ID":"ed7dab29-a923-4ace-811a-cad3279c16ed","Type":"ContainerStarted","Data":"09fa863decd8b43d48b1bb7d548f5e0b2fcece81444c575d7aff50fa57eedc1a"} Dec 04 19:24:05 crc kubenswrapper[4733]: I1204 19:24:05.034543 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2e40-account-create-update-jtgj4" event={"ID":"ef49d872-16fc-4e04-b146-5a7c9b260111","Type":"ContainerStarted","Data":"b85cec2af382613c0ea142e1e44a0ad3fea9714de79af19623cd391e57505a4d"} Dec 04 19:24:05 crc kubenswrapper[4733]: I1204 19:24:05.034601 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2e40-account-create-update-jtgj4" event={"ID":"ef49d872-16fc-4e04-b146-5a7c9b260111","Type":"ContainerStarted","Data":"b194b9c91febd25d7e6292f8e7d4e8d8fcb4cb1c4b82e1802cb409198463812d"} Dec 04 19:24:05 crc kubenswrapper[4733]: I1204 19:24:05.071465 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-2e40-account-create-update-jtgj4" podStartSLOduration=2.071446636 podStartE2EDuration="2.071446636s" podCreationTimestamp="2025-12-04 19:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:24:05.064889019 +0000 UTC m=+6307.020250075" watchObservedRunningTime="2025-12-04 19:24:05.071446636 +0000 UTC m=+6307.026807672" Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.050644 4733 generic.go:334] "Generic (PLEG): container finished" podID="ef49d872-16fc-4e04-b146-5a7c9b260111" containerID="b85cec2af382613c0ea142e1e44a0ad3fea9714de79af19623cd391e57505a4d" exitCode=0 Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.051320 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2e40-account-create-update-jtgj4" event={"ID":"ef49d872-16fc-4e04-b146-5a7c9b260111","Type":"ContainerDied","Data":"b85cec2af382613c0ea142e1e44a0ad3fea9714de79af19623cd391e57505a4d"} Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.457082 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.528173 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b8j4\" (UniqueName: \"kubernetes.io/projected/ed7dab29-a923-4ace-811a-cad3279c16ed-kube-api-access-4b8j4\") pod \"ed7dab29-a923-4ace-811a-cad3279c16ed\" (UID: \"ed7dab29-a923-4ace-811a-cad3279c16ed\") " Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.528259 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7dab29-a923-4ace-811a-cad3279c16ed-operator-scripts\") pod \"ed7dab29-a923-4ace-811a-cad3279c16ed\" (UID: \"ed7dab29-a923-4ace-811a-cad3279c16ed\") " Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.529099 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed7dab29-a923-4ace-811a-cad3279c16ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed7dab29-a923-4ace-811a-cad3279c16ed" (UID: "ed7dab29-a923-4ace-811a-cad3279c16ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.542260 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7dab29-a923-4ace-811a-cad3279c16ed-kube-api-access-4b8j4" (OuterVolumeSpecName: "kube-api-access-4b8j4") pod "ed7dab29-a923-4ace-811a-cad3279c16ed" (UID: "ed7dab29-a923-4ace-811a-cad3279c16ed"). InnerVolumeSpecName "kube-api-access-4b8j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.630522 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7dab29-a923-4ace-811a-cad3279c16ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:06 crc kubenswrapper[4733]: I1204 19:24:06.630571 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b8j4\" (UniqueName: \"kubernetes.io/projected/ed7dab29-a923-4ace-811a-cad3279c16ed-kube-api-access-4b8j4\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.073627 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ft7hm" event={"ID":"ed7dab29-a923-4ace-811a-cad3279c16ed","Type":"ContainerDied","Data":"09fa863decd8b43d48b1bb7d548f5e0b2fcece81444c575d7aff50fa57eedc1a"} Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.073690 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ft7hm" Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.073711 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09fa863decd8b43d48b1bb7d548f5e0b2fcece81444c575d7aff50fa57eedc1a" Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.471451 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.547158 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef49d872-16fc-4e04-b146-5a7c9b260111-operator-scripts\") pod \"ef49d872-16fc-4e04-b146-5a7c9b260111\" (UID: \"ef49d872-16fc-4e04-b146-5a7c9b260111\") " Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.547617 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef49d872-16fc-4e04-b146-5a7c9b260111-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef49d872-16fc-4e04-b146-5a7c9b260111" (UID: "ef49d872-16fc-4e04-b146-5a7c9b260111"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.547704 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k45b4\" (UniqueName: \"kubernetes.io/projected/ef49d872-16fc-4e04-b146-5a7c9b260111-kube-api-access-k45b4\") pod \"ef49d872-16fc-4e04-b146-5a7c9b260111\" (UID: \"ef49d872-16fc-4e04-b146-5a7c9b260111\") " Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.548589 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef49d872-16fc-4e04-b146-5a7c9b260111-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.561073 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef49d872-16fc-4e04-b146-5a7c9b260111-kube-api-access-k45b4" (OuterVolumeSpecName: "kube-api-access-k45b4") pod "ef49d872-16fc-4e04-b146-5a7c9b260111" (UID: "ef49d872-16fc-4e04-b146-5a7c9b260111"). InnerVolumeSpecName "kube-api-access-k45b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:24:07 crc kubenswrapper[4733]: I1204 19:24:07.650677 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k45b4\" (UniqueName: \"kubernetes.io/projected/ef49d872-16fc-4e04-b146-5a7c9b260111-kube-api-access-k45b4\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.096987 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2e40-account-create-update-jtgj4" event={"ID":"ef49d872-16fc-4e04-b146-5a7c9b260111","Type":"ContainerDied","Data":"b194b9c91febd25d7e6292f8e7d4e8d8fcb4cb1c4b82e1802cb409198463812d"} Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.097052 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b194b9c91febd25d7e6292f8e7d4e8d8fcb4cb1c4b82e1802cb409198463812d" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.097150 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2e40-account-create-update-jtgj4" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.825597 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-xvqzk"] Dec 04 19:24:08 crc kubenswrapper[4733]: E1204 19:24:08.826397 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef49d872-16fc-4e04-b146-5a7c9b260111" containerName="mariadb-account-create-update" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.826424 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef49d872-16fc-4e04-b146-5a7c9b260111" containerName="mariadb-account-create-update" Dec 04 19:24:08 crc kubenswrapper[4733]: E1204 19:24:08.826447 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7dab29-a923-4ace-811a-cad3279c16ed" containerName="mariadb-database-create" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.826455 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7dab29-a923-4ace-811a-cad3279c16ed" containerName="mariadb-database-create" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.826688 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef49d872-16fc-4e04-b146-5a7c9b260111" containerName="mariadb-account-create-update" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.826728 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7dab29-a923-4ace-811a-cad3279c16ed" containerName="mariadb-database-create" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.827551 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.830525 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-zvwbn" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.830709 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.843598 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xvqzk"] Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.872990 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-config-data\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.873057 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sgnh\" (UniqueName: \"kubernetes.io/projected/73f7ba50-f386-4612-a44e-ae29091ce6d0-kube-api-access-5sgnh\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.873244 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-combined-ca-bundle\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.974603 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-combined-ca-bundle\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.974677 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-config-data\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.974704 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sgnh\" (UniqueName: \"kubernetes.io/projected/73f7ba50-f386-4612-a44e-ae29091ce6d0-kube-api-access-5sgnh\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.979135 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-config-data\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.979540 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-combined-ca-bundle\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:08 crc kubenswrapper[4733]: I1204 19:24:08.988660 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sgnh\" (UniqueName: \"kubernetes.io/projected/73f7ba50-f386-4612-a44e-ae29091ce6d0-kube-api-access-5sgnh\") pod \"heat-db-sync-xvqzk\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:09 crc kubenswrapper[4733]: I1204 19:24:09.152709 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:09 crc kubenswrapper[4733]: I1204 19:24:09.633495 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xvqzk"] Dec 04 19:24:09 crc kubenswrapper[4733]: W1204 19:24:09.637959 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73f7ba50_f386_4612_a44e_ae29091ce6d0.slice/crio-3b00dc1aec97ab64b2f59ed2337ff680019728c05190fbfe99c0ffa12aa87248 WatchSource:0}: Error finding container 3b00dc1aec97ab64b2f59ed2337ff680019728c05190fbfe99c0ffa12aa87248: Status 404 returned error can't find the container with id 3b00dc1aec97ab64b2f59ed2337ff680019728c05190fbfe99c0ffa12aa87248 Dec 04 19:24:10 crc kubenswrapper[4733]: I1204 19:24:10.118580 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xvqzk" event={"ID":"73f7ba50-f386-4612-a44e-ae29091ce6d0","Type":"ContainerStarted","Data":"3b00dc1aec97ab64b2f59ed2337ff680019728c05190fbfe99c0ffa12aa87248"} Dec 04 19:24:12 crc kubenswrapper[4733]: I1204 19:24:12.692633 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:12 crc kubenswrapper[4733]: I1204 19:24:12.693216 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:17 crc kubenswrapper[4733]: I1204 19:24:17.204080 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xvqzk" event={"ID":"73f7ba50-f386-4612-a44e-ae29091ce6d0","Type":"ContainerStarted","Data":"8f9a3c448cf77fa9c7136d0326b7da058edfb480ff4aae96e94b30255ad5d50f"} Dec 04 19:24:17 crc kubenswrapper[4733]: I1204 19:24:17.224667 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-xvqzk" podStartSLOduration=2.604030334 podStartE2EDuration="9.224643396s" podCreationTimestamp="2025-12-04 19:24:08 +0000 UTC" firstStartedPulling="2025-12-04 19:24:09.640032398 +0000 UTC m=+6311.595393444" lastFinishedPulling="2025-12-04 19:24:16.26064546 +0000 UTC m=+6318.216006506" observedRunningTime="2025-12-04 19:24:17.216555077 +0000 UTC m=+6319.171916133" watchObservedRunningTime="2025-12-04 19:24:17.224643396 +0000 UTC m=+6319.180004442" Dec 04 19:24:19 crc kubenswrapper[4733]: I1204 19:24:19.229171 4733 generic.go:334] "Generic (PLEG): container finished" podID="73f7ba50-f386-4612-a44e-ae29091ce6d0" containerID="8f9a3c448cf77fa9c7136d0326b7da058edfb480ff4aae96e94b30255ad5d50f" exitCode=0 Dec 04 19:24:19 crc kubenswrapper[4733]: I1204 19:24:19.229318 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xvqzk" event={"ID":"73f7ba50-f386-4612-a44e-ae29091ce6d0","Type":"ContainerDied","Data":"8f9a3c448cf77fa9c7136d0326b7da058edfb480ff4aae96e94b30255ad5d50f"} Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.613713 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.616989 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-combined-ca-bundle\") pod \"73f7ba50-f386-4612-a44e-ae29091ce6d0\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.617098 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sgnh\" (UniqueName: \"kubernetes.io/projected/73f7ba50-f386-4612-a44e-ae29091ce6d0-kube-api-access-5sgnh\") pod \"73f7ba50-f386-4612-a44e-ae29091ce6d0\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.617204 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-config-data\") pod \"73f7ba50-f386-4612-a44e-ae29091ce6d0\" (UID: \"73f7ba50-f386-4612-a44e-ae29091ce6d0\") " Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.622378 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f7ba50-f386-4612-a44e-ae29091ce6d0-kube-api-access-5sgnh" (OuterVolumeSpecName: "kube-api-access-5sgnh") pod "73f7ba50-f386-4612-a44e-ae29091ce6d0" (UID: "73f7ba50-f386-4612-a44e-ae29091ce6d0"). InnerVolumeSpecName "kube-api-access-5sgnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.664457 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73f7ba50-f386-4612-a44e-ae29091ce6d0" (UID: "73f7ba50-f386-4612-a44e-ae29091ce6d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.710953 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-config-data" (OuterVolumeSpecName: "config-data") pod "73f7ba50-f386-4612-a44e-ae29091ce6d0" (UID: "73f7ba50-f386-4612-a44e-ae29091ce6d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.721744 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.721845 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sgnh\" (UniqueName: \"kubernetes.io/projected/73f7ba50-f386-4612-a44e-ae29091ce6d0-kube-api-access-5sgnh\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:20 crc kubenswrapper[4733]: I1204 19:24:20.721870 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7ba50-f386-4612-a44e-ae29091ce6d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:21 crc kubenswrapper[4733]: I1204 19:24:21.993295 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xvqzk" event={"ID":"73f7ba50-f386-4612-a44e-ae29091ce6d0","Type":"ContainerDied","Data":"3b00dc1aec97ab64b2f59ed2337ff680019728c05190fbfe99c0ffa12aa87248"} Dec 04 19:24:21 crc kubenswrapper[4733]: I1204 19:24:21.994290 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b00dc1aec97ab64b2f59ed2337ff680019728c05190fbfe99c0ffa12aa87248" Dec 04 19:24:21 crc kubenswrapper[4733]: I1204 19:24:21.994447 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xvqzk" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.025977 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-68ff895d44-vq5jt"] Dec 04 19:24:23 crc kubenswrapper[4733]: E1204 19:24:23.027760 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f7ba50-f386-4612-a44e-ae29091ce6d0" containerName="heat-db-sync" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.027884 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f7ba50-f386-4612-a44e-ae29091ce6d0" containerName="heat-db-sync" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.028346 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f7ba50-f386-4612-a44e-ae29091ce6d0" containerName="heat-db-sync" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.029127 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.032246 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.032645 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.034724 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-zvwbn" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.057561 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-68ff895d44-vq5jt"] Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.190961 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-config-data-custom\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.191032 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-config-data\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.191205 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2p4t\" (UniqueName: \"kubernetes.io/projected/808a8006-fe18-43ee-b546-4a6043eb1081-kube-api-access-m2p4t\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.191264 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-combined-ca-bundle\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.229089 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-686759f4c9-dw5h2"] Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.231123 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.235305 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.249990 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-686759f4c9-dw5h2"] Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.272270 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-66c8bffccb-r24sl"] Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.273580 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.279903 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.283940 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66c8bffccb-r24sl"] Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292278 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-config-data-custom\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292315 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tljhq\" (UniqueName: \"kubernetes.io/projected/865712a6-df7e-486a-9cae-bd0a0de5556a-kube-api-access-tljhq\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292364 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-config-data\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292388 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-config-data\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292408 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-config-data-custom\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292441 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2p4t\" (UniqueName: \"kubernetes.io/projected/808a8006-fe18-43ee-b546-4a6043eb1081-kube-api-access-m2p4t\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292461 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-combined-ca-bundle\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292486 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-config-data\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292509 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-config-data-custom\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292529 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bzvc\" (UniqueName: \"kubernetes.io/projected/37ac4820-fee5-490a-83be-48d01d292e8f-kube-api-access-2bzvc\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292549 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-combined-ca-bundle\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.292575 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-combined-ca-bundle\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.373777 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-config-data-custom\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.379255 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2p4t\" (UniqueName: \"kubernetes.io/projected/808a8006-fe18-43ee-b546-4a6043eb1081-kube-api-access-m2p4t\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.379499 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-config-data\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.380348 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/808a8006-fe18-43ee-b546-4a6043eb1081-combined-ca-bundle\") pod \"heat-engine-68ff895d44-vq5jt\" (UID: \"808a8006-fe18-43ee-b546-4a6043eb1081\") " pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.397058 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-combined-ca-bundle\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.397202 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-config-data\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.397269 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-config-data-custom\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.397317 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bzvc\" (UniqueName: \"kubernetes.io/projected/37ac4820-fee5-490a-83be-48d01d292e8f-kube-api-access-2bzvc\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.397759 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-combined-ca-bundle\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.398007 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tljhq\" (UniqueName: \"kubernetes.io/projected/865712a6-df7e-486a-9cae-bd0a0de5556a-kube-api-access-tljhq\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.398189 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-config-data\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.398253 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-config-data-custom\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.402207 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-combined-ca-bundle\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.404582 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-combined-ca-bundle\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.405907 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-config-data-custom\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.408126 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37ac4820-fee5-490a-83be-48d01d292e8f-config-data\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.419300 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bzvc\" (UniqueName: \"kubernetes.io/projected/37ac4820-fee5-490a-83be-48d01d292e8f-kube-api-access-2bzvc\") pod \"heat-cfnapi-66c8bffccb-r24sl\" (UID: \"37ac4820-fee5-490a-83be-48d01d292e8f\") " pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.419590 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-config-data-custom\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.419913 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865712a6-df7e-486a-9cae-bd0a0de5556a-config-data\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.424727 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tljhq\" (UniqueName: \"kubernetes.io/projected/865712a6-df7e-486a-9cae-bd0a0de5556a-kube-api-access-tljhq\") pod \"heat-api-686759f4c9-dw5h2\" (UID: \"865712a6-df7e-486a-9cae-bd0a0de5556a\") " pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.573332 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.651392 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.663491 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.732776 4733 scope.go:117] "RemoveContainer" containerID="d86ee3847eee51d74f52491ec6b66e0b770d53015dd77d5e3e38986499bafac9" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.838430 4733 scope.go:117] "RemoveContainer" containerID="a60c6dc6ddfd19c65af9c8a9b2d5c196b947bfe4a4e103adef76e8f69a632977" Dec 04 19:24:23 crc kubenswrapper[4733]: I1204 19:24:23.911060 4733 scope.go:117] "RemoveContainer" containerID="30be18b9098e9150bac2d3e92ff313a66b5f42c5c7df82fcd9b06eac20686d81" Dec 04 19:24:24 crc kubenswrapper[4733]: I1204 19:24:24.160265 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-686759f4c9-dw5h2"] Dec 04 19:24:24 crc kubenswrapper[4733]: I1204 19:24:24.273762 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-68ff895d44-vq5jt"] Dec 04 19:24:24 crc kubenswrapper[4733]: W1204 19:24:24.279735 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod808a8006_fe18_43ee_b546_4a6043eb1081.slice/crio-bbfb49cb55dcbb1e4cfc1957ce6c8d59999db44163ece7bf84f0009d336e9170 WatchSource:0}: Error finding container bbfb49cb55dcbb1e4cfc1957ce6c8d59999db44163ece7bf84f0009d336e9170: Status 404 returned error can't find the container with id bbfb49cb55dcbb1e4cfc1957ce6c8d59999db44163ece7bf84f0009d336e9170 Dec 04 19:24:24 crc kubenswrapper[4733]: I1204 19:24:24.403523 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66c8bffccb-r24sl"] Dec 04 19:24:24 crc kubenswrapper[4733]: W1204 19:24:24.412550 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37ac4820_fee5_490a_83be_48d01d292e8f.slice/crio-4e068543aa67ad58e5d6de4182d45d1cd345d6889506f68b1d55428bc3adb0c0 WatchSource:0}: Error finding container 4e068543aa67ad58e5d6de4182d45d1cd345d6889506f68b1d55428bc3adb0c0: Status 404 returned error can't find the container with id 4e068543aa67ad58e5d6de4182d45d1cd345d6889506f68b1d55428bc3adb0c0 Dec 04 19:24:25 crc kubenswrapper[4733]: I1204 19:24:25.031061 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66c8bffccb-r24sl" event={"ID":"37ac4820-fee5-490a-83be-48d01d292e8f","Type":"ContainerStarted","Data":"4e068543aa67ad58e5d6de4182d45d1cd345d6889506f68b1d55428bc3adb0c0"} Dec 04 19:24:25 crc kubenswrapper[4733]: I1204 19:24:25.040131 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-686759f4c9-dw5h2" event={"ID":"865712a6-df7e-486a-9cae-bd0a0de5556a","Type":"ContainerStarted","Data":"a0725572f789c401d2f0fe55a0c8ff5cb3277e83b44ed1fed0a52ffb4f206f96"} Dec 04 19:24:25 crc kubenswrapper[4733]: I1204 19:24:25.042986 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-68ff895d44-vq5jt" event={"ID":"808a8006-fe18-43ee-b546-4a6043eb1081","Type":"ContainerStarted","Data":"0518181c73931376093c3b86fc90e40fc001be630ce95fb37024067416d31ca1"} Dec 04 19:24:25 crc kubenswrapper[4733]: I1204 19:24:25.043032 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-68ff895d44-vq5jt" event={"ID":"808a8006-fe18-43ee-b546-4a6043eb1081","Type":"ContainerStarted","Data":"bbfb49cb55dcbb1e4cfc1957ce6c8d59999db44163ece7bf84f0009d336e9170"} Dec 04 19:24:25 crc kubenswrapper[4733]: I1204 19:24:25.044126 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:25 crc kubenswrapper[4733]: I1204 19:24:25.060815 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:25 crc kubenswrapper[4733]: I1204 19:24:25.061463 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-68ff895d44-vq5jt" podStartSLOduration=2.061443533 podStartE2EDuration="2.061443533s" podCreationTimestamp="2025-12-04 19:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:24:25.05871709 +0000 UTC m=+6327.014078126" watchObservedRunningTime="2025-12-04 19:24:25.061443533 +0000 UTC m=+6327.016804579" Dec 04 19:24:27 crc kubenswrapper[4733]: I1204 19:24:27.140565 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5884fc67b7-6bjb4" Dec 04 19:24:27 crc kubenswrapper[4733]: I1204 19:24:27.213383 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66d7c5c989-8t52p"] Dec 04 19:24:27 crc kubenswrapper[4733]: I1204 19:24:27.213676 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66d7c5c989-8t52p" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon-log" containerID="cri-o://0b69c74c8f7594a7cc51e4de12f8e94fcfe817a7afa5f0aee1f4be5669d9bb32" gracePeriod=30 Dec 04 19:24:27 crc kubenswrapper[4733]: I1204 19:24:27.214185 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66d7c5c989-8t52p" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon" containerID="cri-o://63f0449d7ee89d9ebb20cd215985fea362dafbff08e29ebb67213c03e5e2846c" gracePeriod=30 Dec 04 19:24:28 crc kubenswrapper[4733]: I1204 19:24:28.072729 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66c8bffccb-r24sl" event={"ID":"37ac4820-fee5-490a-83be-48d01d292e8f","Type":"ContainerStarted","Data":"d97ed7e648c81825504ba0e0e6d79b0147f3b3aa391eac81f8b29ee83559e18f"} Dec 04 19:24:28 crc kubenswrapper[4733]: I1204 19:24:28.073132 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:28 crc kubenswrapper[4733]: I1204 19:24:28.075122 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-686759f4c9-dw5h2" event={"ID":"865712a6-df7e-486a-9cae-bd0a0de5556a","Type":"ContainerStarted","Data":"85871feda29ce83505197869252e9df39e64aa332a4570b4f9e0b3a47f6b61b3"} Dec 04 19:24:28 crc kubenswrapper[4733]: I1204 19:24:28.075569 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:28 crc kubenswrapper[4733]: I1204 19:24:28.090386 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-66c8bffccb-r24sl" podStartSLOduration=1.834203078 podStartE2EDuration="5.090373019s" podCreationTimestamp="2025-12-04 19:24:23 +0000 UTC" firstStartedPulling="2025-12-04 19:24:24.417131975 +0000 UTC m=+6326.372493021" lastFinishedPulling="2025-12-04 19:24:27.673301916 +0000 UTC m=+6329.628662962" observedRunningTime="2025-12-04 19:24:28.087648476 +0000 UTC m=+6330.043009532" watchObservedRunningTime="2025-12-04 19:24:28.090373019 +0000 UTC m=+6330.045734065" Dec 04 19:24:28 crc kubenswrapper[4733]: I1204 19:24:28.109374 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-686759f4c9-dw5h2" podStartSLOduration=1.608340683 podStartE2EDuration="5.109356964s" podCreationTimestamp="2025-12-04 19:24:23 +0000 UTC" firstStartedPulling="2025-12-04 19:24:24.165695526 +0000 UTC m=+6326.121056572" lastFinishedPulling="2025-12-04 19:24:27.666711797 +0000 UTC m=+6329.622072853" observedRunningTime="2025-12-04 19:24:28.105555701 +0000 UTC m=+6330.060916747" watchObservedRunningTime="2025-12-04 19:24:28.109356964 +0000 UTC m=+6330.064718010" Dec 04 19:24:31 crc kubenswrapper[4733]: I1204 19:24:31.112549 4733 generic.go:334] "Generic (PLEG): container finished" podID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerID="63f0449d7ee89d9ebb20cd215985fea362dafbff08e29ebb67213c03e5e2846c" exitCode=0 Dec 04 19:24:31 crc kubenswrapper[4733]: I1204 19:24:31.112618 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d7c5c989-8t52p" event={"ID":"8f339b4d-9773-4b73-8828-7cbe31882ff7","Type":"ContainerDied","Data":"63f0449d7ee89d9ebb20cd215985fea362dafbff08e29ebb67213c03e5e2846c"} Dec 04 19:24:33 crc kubenswrapper[4733]: I1204 19:24:33.053087 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-8wgfz"] Dec 04 19:24:33 crc kubenswrapper[4733]: I1204 19:24:33.077500 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-fd5b-account-create-update-pmmtr"] Dec 04 19:24:33 crc kubenswrapper[4733]: I1204 19:24:33.093236 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-8wgfz"] Dec 04 19:24:33 crc kubenswrapper[4733]: I1204 19:24:33.104341 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-fd5b-account-create-update-pmmtr"] Dec 04 19:24:34 crc kubenswrapper[4733]: I1204 19:24:34.348224 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9643222a-15e6-40a3-b978-68e777c387d5" path="/var/lib/kubelet/pods/9643222a-15e6-40a3-b978-68e777c387d5/volumes" Dec 04 19:24:34 crc kubenswrapper[4733]: I1204 19:24:34.350288 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1" path="/var/lib/kubelet/pods/e6fd0f2d-3c8e-4ab2-892f-d475d6b58af1/volumes" Dec 04 19:24:34 crc kubenswrapper[4733]: I1204 19:24:34.892140 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-686759f4c9-dw5h2" Dec 04 19:24:35 crc kubenswrapper[4733]: I1204 19:24:35.466331 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-66c8bffccb-r24sl" Dec 04 19:24:35 crc kubenswrapper[4733]: I1204 19:24:35.894503 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66d7c5c989-8t52p" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.117:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8080: connect: connection refused" Dec 04 19:24:39 crc kubenswrapper[4733]: I1204 19:24:39.036632 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-bdqb6"] Dec 04 19:24:39 crc kubenswrapper[4733]: I1204 19:24:39.046065 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-bdqb6"] Dec 04 19:24:40 crc kubenswrapper[4733]: I1204 19:24:40.349533 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10" path="/var/lib/kubelet/pods/2fe62d0c-ea33-47e8-9cf0-f81a9bbaee10/volumes" Dec 04 19:24:43 crc kubenswrapper[4733]: I1204 19:24:43.692498 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-68ff895d44-vq5jt" Dec 04 19:24:45 crc kubenswrapper[4733]: I1204 19:24:45.894580 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66d7c5c989-8t52p" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.117:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8080: connect: connection refused" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.496258 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s"] Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.505605 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.513582 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.524018 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s"] Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.575030 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.575076 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.575116 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq6n9\" (UniqueName: \"kubernetes.io/projected/d4853046-f3d5-48dc-8bb8-aa019bf731dd-kube-api-access-sq6n9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.677036 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.677099 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.677146 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq6n9\" (UniqueName: \"kubernetes.io/projected/d4853046-f3d5-48dc-8bb8-aa019bf731dd-kube-api-access-sq6n9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.677672 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.677941 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.701925 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq6n9\" (UniqueName: \"kubernetes.io/projected/d4853046-f3d5-48dc-8bb8-aa019bf731dd-kube-api-access-sq6n9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:54 crc kubenswrapper[4733]: I1204 19:24:54.842185 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:24:55 crc kubenswrapper[4733]: I1204 19:24:55.346239 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s"] Dec 04 19:24:55 crc kubenswrapper[4733]: I1204 19:24:55.355572 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" event={"ID":"d4853046-f3d5-48dc-8bb8-aa019bf731dd","Type":"ContainerStarted","Data":"f932f476cf8c4126c2b2f5e3c104c2e0d436f5257673b0eb176e04d40ba6310d"} Dec 04 19:24:55 crc kubenswrapper[4733]: I1204 19:24:55.894682 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66d7c5c989-8t52p" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.117:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8080: connect: connection refused" Dec 04 19:24:55 crc kubenswrapper[4733]: I1204 19:24:55.895262 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:24:56 crc kubenswrapper[4733]: I1204 19:24:56.364882 4733 generic.go:334] "Generic (PLEG): container finished" podID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerID="ed9fc05e1102d7897c8ad37c13fc1068c278cc3724369e2d0ac28b98d4fb322d" exitCode=0 Dec 04 19:24:56 crc kubenswrapper[4733]: I1204 19:24:56.364933 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" event={"ID":"d4853046-f3d5-48dc-8bb8-aa019bf731dd","Type":"ContainerDied","Data":"ed9fc05e1102d7897c8ad37c13fc1068c278cc3724369e2d0ac28b98d4fb322d"} Dec 04 19:24:56 crc kubenswrapper[4733]: I1204 19:24:56.366988 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.382014 4733 generic.go:334] "Generic (PLEG): container finished" podID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerID="0b69c74c8f7594a7cc51e4de12f8e94fcfe817a7afa5f0aee1f4be5669d9bb32" exitCode=137 Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.382111 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d7c5c989-8t52p" event={"ID":"8f339b4d-9773-4b73-8828-7cbe31882ff7","Type":"ContainerDied","Data":"0b69c74c8f7594a7cc51e4de12f8e94fcfe817a7afa5f0aee1f4be5669d9bb32"} Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.856039 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.955258 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f339b4d-9773-4b73-8828-7cbe31882ff7-horizon-secret-key\") pod \"8f339b4d-9773-4b73-8828-7cbe31882ff7\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.955386 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f339b4d-9773-4b73-8828-7cbe31882ff7-logs\") pod \"8f339b4d-9773-4b73-8828-7cbe31882ff7\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.955490 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-scripts\") pod \"8f339b4d-9773-4b73-8828-7cbe31882ff7\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.955550 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6w66\" (UniqueName: \"kubernetes.io/projected/8f339b4d-9773-4b73-8828-7cbe31882ff7-kube-api-access-n6w66\") pod \"8f339b4d-9773-4b73-8828-7cbe31882ff7\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.955624 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-config-data\") pod \"8f339b4d-9773-4b73-8828-7cbe31882ff7\" (UID: \"8f339b4d-9773-4b73-8828-7cbe31882ff7\") " Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.955916 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f339b4d-9773-4b73-8828-7cbe31882ff7-logs" (OuterVolumeSpecName: "logs") pod "8f339b4d-9773-4b73-8828-7cbe31882ff7" (UID: "8f339b4d-9773-4b73-8828-7cbe31882ff7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.956069 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f339b4d-9773-4b73-8828-7cbe31882ff7-logs\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.965443 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f339b4d-9773-4b73-8828-7cbe31882ff7-kube-api-access-n6w66" (OuterVolumeSpecName: "kube-api-access-n6w66") pod "8f339b4d-9773-4b73-8828-7cbe31882ff7" (UID: "8f339b4d-9773-4b73-8828-7cbe31882ff7"). InnerVolumeSpecName "kube-api-access-n6w66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.965503 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f339b4d-9773-4b73-8828-7cbe31882ff7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8f339b4d-9773-4b73-8828-7cbe31882ff7" (UID: "8f339b4d-9773-4b73-8828-7cbe31882ff7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:24:57 crc kubenswrapper[4733]: I1204 19:24:57.991527 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-config-data" (OuterVolumeSpecName: "config-data") pod "8f339b4d-9773-4b73-8828-7cbe31882ff7" (UID: "8f339b4d-9773-4b73-8828-7cbe31882ff7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.008636 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-scripts" (OuterVolumeSpecName: "scripts") pod "8f339b4d-9773-4b73-8828-7cbe31882ff7" (UID: "8f339b4d-9773-4b73-8828-7cbe31882ff7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.058106 4733 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f339b4d-9773-4b73-8828-7cbe31882ff7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.058138 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.058148 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6w66\" (UniqueName: \"kubernetes.io/projected/8f339b4d-9773-4b73-8828-7cbe31882ff7-kube-api-access-n6w66\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.058160 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f339b4d-9773-4b73-8828-7cbe31882ff7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.402593 4733 generic.go:334] "Generic (PLEG): container finished" podID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerID="d4115a825d3d28c7134c4974f5ed5b46ce638a711c98c324cfaa4d567810143c" exitCode=0 Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.402688 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" event={"ID":"d4853046-f3d5-48dc-8bb8-aa019bf731dd","Type":"ContainerDied","Data":"d4115a825d3d28c7134c4974f5ed5b46ce638a711c98c324cfaa4d567810143c"} Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.411009 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66d7c5c989-8t52p" event={"ID":"8f339b4d-9773-4b73-8828-7cbe31882ff7","Type":"ContainerDied","Data":"fe38323aa3e661b6285e056850d733fecd56bf524b0c2ac282c17e48d5c8ace5"} Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.411070 4733 scope.go:117] "RemoveContainer" containerID="63f0449d7ee89d9ebb20cd215985fea362dafbff08e29ebb67213c03e5e2846c" Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.411221 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66d7c5c989-8t52p" Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.456531 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66d7c5c989-8t52p"] Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.464322 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66d7c5c989-8t52p"] Dec 04 19:24:58 crc kubenswrapper[4733]: I1204 19:24:58.573533 4733 scope.go:117] "RemoveContainer" containerID="0b69c74c8f7594a7cc51e4de12f8e94fcfe817a7afa5f0aee1f4be5669d9bb32" Dec 04 19:24:59 crc kubenswrapper[4733]: I1204 19:24:59.422732 4733 generic.go:334] "Generic (PLEG): container finished" podID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerID="dcfeb8758824d9e0eb7a5ef6f382e87177c164e1f631bc058c26e7c8aba34853" exitCode=0 Dec 04 19:24:59 crc kubenswrapper[4733]: I1204 19:24:59.422854 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" event={"ID":"d4853046-f3d5-48dc-8bb8-aa019bf731dd","Type":"ContainerDied","Data":"dcfeb8758824d9e0eb7a5ef6f382e87177c164e1f631bc058c26e7c8aba34853"} Dec 04 19:25:00 crc kubenswrapper[4733]: I1204 19:25:00.350995 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" path="/var/lib/kubelet/pods/8f339b4d-9773-4b73-8828-7cbe31882ff7/volumes" Dec 04 19:25:00 crc kubenswrapper[4733]: I1204 19:25:00.886769 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.019487 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-bundle\") pod \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.019714 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-util\") pod \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.019743 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq6n9\" (UniqueName: \"kubernetes.io/projected/d4853046-f3d5-48dc-8bb8-aa019bf731dd-kube-api-access-sq6n9\") pod \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\" (UID: \"d4853046-f3d5-48dc-8bb8-aa019bf731dd\") " Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.021769 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-bundle" (OuterVolumeSpecName: "bundle") pod "d4853046-f3d5-48dc-8bb8-aa019bf731dd" (UID: "d4853046-f3d5-48dc-8bb8-aa019bf731dd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.026298 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4853046-f3d5-48dc-8bb8-aa019bf731dd-kube-api-access-sq6n9" (OuterVolumeSpecName: "kube-api-access-sq6n9") pod "d4853046-f3d5-48dc-8bb8-aa019bf731dd" (UID: "d4853046-f3d5-48dc-8bb8-aa019bf731dd"). InnerVolumeSpecName "kube-api-access-sq6n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.045143 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-util" (OuterVolumeSpecName: "util") pod "d4853046-f3d5-48dc-8bb8-aa019bf731dd" (UID: "d4853046-f3d5-48dc-8bb8-aa019bf731dd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.121844 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-util\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.122057 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq6n9\" (UniqueName: \"kubernetes.io/projected/d4853046-f3d5-48dc-8bb8-aa019bf731dd-kube-api-access-sq6n9\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.122158 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4853046-f3d5-48dc-8bb8-aa019bf731dd-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.457941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" event={"ID":"d4853046-f3d5-48dc-8bb8-aa019bf731dd","Type":"ContainerDied","Data":"f932f476cf8c4126c2b2f5e3c104c2e0d436f5257673b0eb176e04d40ba6310d"} Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.458248 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f932f476cf8c4126c2b2f5e3c104c2e0d436f5257673b0eb176e04d40ba6310d" Dec 04 19:25:01 crc kubenswrapper[4733]: I1204 19:25:01.458071 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066135 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-55gp7"] Dec 04 19:25:02 crc kubenswrapper[4733]: E1204 19:25:02.066595 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerName="util" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066609 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerName="util" Dec 04 19:25:02 crc kubenswrapper[4733]: E1204 19:25:02.066625 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerName="extract" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066631 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerName="extract" Dec 04 19:25:02 crc kubenswrapper[4733]: E1204 19:25:02.066648 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerName="pull" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066654 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerName="pull" Dec 04 19:25:02 crc kubenswrapper[4733]: E1204 19:25:02.066688 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066696 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon" Dec 04 19:25:02 crc kubenswrapper[4733]: E1204 19:25:02.066712 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon-log" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066720 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon-log" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066938 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon-log" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066967 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4853046-f3d5-48dc-8bb8-aa019bf731dd" containerName="extract" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.066989 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f339b4d-9773-4b73-8828-7cbe31882ff7" containerName="horizon" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.072765 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.112934 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-55gp7"] Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.158372 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjm7z\" (UniqueName: \"kubernetes.io/projected/a8e002d8-550d-4ebc-8a15-52a801602065-kube-api-access-gjm7z\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.158524 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-catalog-content\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.158557 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-utilities\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.260678 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjm7z\" (UniqueName: \"kubernetes.io/projected/a8e002d8-550d-4ebc-8a15-52a801602065-kube-api-access-gjm7z\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.260815 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-catalog-content\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.260856 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-utilities\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.261304 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-utilities\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.261323 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-catalog-content\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.277543 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjm7z\" (UniqueName: \"kubernetes.io/projected/a8e002d8-550d-4ebc-8a15-52a801602065-kube-api-access-gjm7z\") pod \"redhat-marketplace-55gp7\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.418196 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:02 crc kubenswrapper[4733]: I1204 19:25:02.952563 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-55gp7"] Dec 04 19:25:03 crc kubenswrapper[4733]: I1204 19:25:03.482488 4733 generic.go:334] "Generic (PLEG): container finished" podID="a8e002d8-550d-4ebc-8a15-52a801602065" containerID="93f26c35ee80266acc3eaeaf3ed182b6c15dda22a289cad403566324f2527d42" exitCode=0 Dec 04 19:25:03 crc kubenswrapper[4733]: I1204 19:25:03.482723 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55gp7" event={"ID":"a8e002d8-550d-4ebc-8a15-52a801602065","Type":"ContainerDied","Data":"93f26c35ee80266acc3eaeaf3ed182b6c15dda22a289cad403566324f2527d42"} Dec 04 19:25:03 crc kubenswrapper[4733]: I1204 19:25:03.482855 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55gp7" event={"ID":"a8e002d8-550d-4ebc-8a15-52a801602065","Type":"ContainerStarted","Data":"5274700e19e54b4f7980fda6988695097975e733a3cef88107e09fa04a6bd581"} Dec 04 19:25:04 crc kubenswrapper[4733]: I1204 19:25:04.498232 4733 generic.go:334] "Generic (PLEG): container finished" podID="a8e002d8-550d-4ebc-8a15-52a801602065" containerID="eb9642d96d281fe97053ec79698b610d15dcd29c3f5ac52bca00a43a0550aeab" exitCode=0 Dec 04 19:25:04 crc kubenswrapper[4733]: I1204 19:25:04.498275 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55gp7" event={"ID":"a8e002d8-550d-4ebc-8a15-52a801602065","Type":"ContainerDied","Data":"eb9642d96d281fe97053ec79698b610d15dcd29c3f5ac52bca00a43a0550aeab"} Dec 04 19:25:04 crc kubenswrapper[4733]: E1204 19:25:04.575978 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8e002d8_550d_4ebc_8a15_52a801602065.slice/crio-eb9642d96d281fe97053ec79698b610d15dcd29c3f5ac52bca00a43a0550aeab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8e002d8_550d_4ebc_8a15_52a801602065.slice/crio-conmon-eb9642d96d281fe97053ec79698b610d15dcd29c3f5ac52bca00a43a0550aeab.scope\": RecentStats: unable to find data in memory cache]" Dec 04 19:25:05 crc kubenswrapper[4733]: I1204 19:25:05.510563 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55gp7" event={"ID":"a8e002d8-550d-4ebc-8a15-52a801602065","Type":"ContainerStarted","Data":"2102c1c79943dfe4d112d370782cf5152d46286fe6862bc1f0094a2a7ff235f5"} Dec 04 19:25:05 crc kubenswrapper[4733]: I1204 19:25:05.532997 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-55gp7" podStartSLOduration=2.02632285 podStartE2EDuration="3.532980462s" podCreationTimestamp="2025-12-04 19:25:02 +0000 UTC" firstStartedPulling="2025-12-04 19:25:03.488473165 +0000 UTC m=+6365.443834201" lastFinishedPulling="2025-12-04 19:25:04.995130767 +0000 UTC m=+6366.950491813" observedRunningTime="2025-12-04 19:25:05.528268424 +0000 UTC m=+6367.483629480" watchObservedRunningTime="2025-12-04 19:25:05.532980462 +0000 UTC m=+6367.488341508" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.213974 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.216730 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.230762 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.231311 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-l8vvp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.231939 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.237919 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.276311 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.285328 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.289109 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-qp5sn" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.289359 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.297295 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.308983 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.310339 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.333619 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.371141 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnf6w\" (UniqueName: \"kubernetes.io/projected/1aa893e4-e63a-4f24-97bf-a0c41cf27c64-kube-api-access-qnf6w\") pod \"obo-prometheus-operator-668cf9dfbb-5lv7r\" (UID: \"1aa893e4-e63a-4f24-97bf-a0c41cf27c64\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.405905 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-pjs6d"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.407503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.415377 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-vpjl9" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.415664 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.430926 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-pjs6d"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.473473 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f54ccac-c4f7-47fe-b071-87d8c3231848-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp\" (UID: \"7f54ccac-c4f7-47fe-b071-87d8c3231848\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.473537 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37deee19-62e0-46de-bf27-99748ce7a98b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q\" (UID: \"37deee19-62e0-46de-bf27-99748ce7a98b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.473584 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnf6w\" (UniqueName: \"kubernetes.io/projected/1aa893e4-e63a-4f24-97bf-a0c41cf27c64-kube-api-access-qnf6w\") pod \"obo-prometheus-operator-668cf9dfbb-5lv7r\" (UID: \"1aa893e4-e63a-4f24-97bf-a0c41cf27c64\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.473637 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37deee19-62e0-46de-bf27-99748ce7a98b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q\" (UID: \"37deee19-62e0-46de-bf27-99748ce7a98b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.473684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f54ccac-c4f7-47fe-b071-87d8c3231848-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp\" (UID: \"7f54ccac-c4f7-47fe-b071-87d8c3231848\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.494457 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnf6w\" (UniqueName: \"kubernetes.io/projected/1aa893e4-e63a-4f24-97bf-a0c41cf27c64-kube-api-access-qnf6w\") pod \"obo-prometheus-operator-668cf9dfbb-5lv7r\" (UID: \"1aa893e4-e63a-4f24-97bf-a0c41cf27c64\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.576012 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f54ccac-c4f7-47fe-b071-87d8c3231848-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp\" (UID: \"7f54ccac-c4f7-47fe-b071-87d8c3231848\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.576085 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/97bb703c-e296-4b6d-b878-659c7149d970-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-pjs6d\" (UID: \"97bb703c-e296-4b6d-b878-659c7149d970\") " pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.576116 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37deee19-62e0-46de-bf27-99748ce7a98b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q\" (UID: \"37deee19-62e0-46de-bf27-99748ce7a98b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.576220 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37deee19-62e0-46de-bf27-99748ce7a98b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q\" (UID: \"37deee19-62e0-46de-bf27-99748ce7a98b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.576260 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f54ccac-c4f7-47fe-b071-87d8c3231848-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp\" (UID: \"7f54ccac-c4f7-47fe-b071-87d8c3231848\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.576307 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slfrl\" (UniqueName: \"kubernetes.io/projected/97bb703c-e296-4b6d-b878-659c7149d970-kube-api-access-slfrl\") pod \"observability-operator-d8bb48f5d-pjs6d\" (UID: \"97bb703c-e296-4b6d-b878-659c7149d970\") " pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.595362 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.595714 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f54ccac-c4f7-47fe-b071-87d8c3231848-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp\" (UID: \"7f54ccac-c4f7-47fe-b071-87d8c3231848\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.595742 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37deee19-62e0-46de-bf27-99748ce7a98b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q\" (UID: \"37deee19-62e0-46de-bf27-99748ce7a98b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.595785 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37deee19-62e0-46de-bf27-99748ce7a98b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q\" (UID: \"37deee19-62e0-46de-bf27-99748ce7a98b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.596148 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f54ccac-c4f7-47fe-b071-87d8c3231848-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp\" (UID: \"7f54ccac-c4f7-47fe-b071-87d8c3231848\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.604312 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-95nbh"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.605656 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.607100 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-6f7vm" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.609824 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.637276 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.643929 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-95nbh"] Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.681228 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/97bb703c-e296-4b6d-b878-659c7149d970-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-pjs6d\" (UID: \"97bb703c-e296-4b6d-b878-659c7149d970\") " pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.681518 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slfrl\" (UniqueName: \"kubernetes.io/projected/97bb703c-e296-4b6d-b878-659c7149d970-kube-api-access-slfrl\") pod \"observability-operator-d8bb48f5d-pjs6d\" (UID: \"97bb703c-e296-4b6d-b878-659c7149d970\") " pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.697671 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/97bb703c-e296-4b6d-b878-659c7149d970-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-pjs6d\" (UID: \"97bb703c-e296-4b6d-b878-659c7149d970\") " pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.720185 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slfrl\" (UniqueName: \"kubernetes.io/projected/97bb703c-e296-4b6d-b878-659c7149d970-kube-api-access-slfrl\") pod \"observability-operator-d8bb48f5d-pjs6d\" (UID: \"97bb703c-e296-4b6d-b878-659c7149d970\") " pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.742968 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.783271 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9ldx\" (UniqueName: \"kubernetes.io/projected/3e69f6b5-cf40-4ebe-9949-71031eb10fc8-kube-api-access-g9ldx\") pod \"perses-operator-5446b9c989-95nbh\" (UID: \"3e69f6b5-cf40-4ebe-9949-71031eb10fc8\") " pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.783353 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e69f6b5-cf40-4ebe-9949-71031eb10fc8-openshift-service-ca\") pod \"perses-operator-5446b9c989-95nbh\" (UID: \"3e69f6b5-cf40-4ebe-9949-71031eb10fc8\") " pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.886054 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9ldx\" (UniqueName: \"kubernetes.io/projected/3e69f6b5-cf40-4ebe-9949-71031eb10fc8-kube-api-access-g9ldx\") pod \"perses-operator-5446b9c989-95nbh\" (UID: \"3e69f6b5-cf40-4ebe-9949-71031eb10fc8\") " pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.886114 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e69f6b5-cf40-4ebe-9949-71031eb10fc8-openshift-service-ca\") pod \"perses-operator-5446b9c989-95nbh\" (UID: \"3e69f6b5-cf40-4ebe-9949-71031eb10fc8\") " pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.887418 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e69f6b5-cf40-4ebe-9949-71031eb10fc8-openshift-service-ca\") pod \"perses-operator-5446b9c989-95nbh\" (UID: \"3e69f6b5-cf40-4ebe-9949-71031eb10fc8\") " pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:11 crc kubenswrapper[4733]: I1204 19:25:11.927497 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9ldx\" (UniqueName: \"kubernetes.io/projected/3e69f6b5-cf40-4ebe-9949-71031eb10fc8-kube-api-access-g9ldx\") pod \"perses-operator-5446b9c989-95nbh\" (UID: \"3e69f6b5-cf40-4ebe-9949-71031eb10fc8\") " pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.070875 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.423374 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.423410 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.477297 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.664416 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.793486 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp"] Dec 04 19:25:12 crc kubenswrapper[4733]: W1204 19:25:12.794958 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37deee19_62e0_46de_bf27_99748ce7a98b.slice/crio-7fb63489046df50304c398612d5bd3891c86cc09a220acb0da1b338f3fb43e02 WatchSource:0}: Error finding container 7fb63489046df50304c398612d5bd3891c86cc09a220acb0da1b338f3fb43e02: Status 404 returned error can't find the container with id 7fb63489046df50304c398612d5bd3891c86cc09a220acb0da1b338f3fb43e02 Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.806428 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r"] Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.815976 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q"] Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.832325 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-pjs6d"] Dec 04 19:25:12 crc kubenswrapper[4733]: I1204 19:25:12.952331 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-95nbh"] Dec 04 19:25:13 crc kubenswrapper[4733]: I1204 19:25:13.587944 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" event={"ID":"97bb703c-e296-4b6d-b878-659c7149d970","Type":"ContainerStarted","Data":"cca21e63e5f6a49707ac961e0edc49e8619c58233985090307cdeb6e8a0e4f6e"} Dec 04 19:25:13 crc kubenswrapper[4733]: I1204 19:25:13.589057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-95nbh" event={"ID":"3e69f6b5-cf40-4ebe-9949-71031eb10fc8","Type":"ContainerStarted","Data":"5b2b282a208c9949b6fecd1addfd0355f06f5d884554e42c8210b872858dc4fa"} Dec 04 19:25:13 crc kubenswrapper[4733]: I1204 19:25:13.590085 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" event={"ID":"37deee19-62e0-46de-bf27-99748ce7a98b","Type":"ContainerStarted","Data":"7fb63489046df50304c398612d5bd3891c86cc09a220acb0da1b338f3fb43e02"} Dec 04 19:25:13 crc kubenswrapper[4733]: I1204 19:25:13.591183 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" event={"ID":"7f54ccac-c4f7-47fe-b071-87d8c3231848","Type":"ContainerStarted","Data":"6abada2528f288d7cfdc7e56f99e6155fd43a158b16526c94fae559478724592"} Dec 04 19:25:13 crc kubenswrapper[4733]: I1204 19:25:13.593148 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r" event={"ID":"1aa893e4-e63a-4f24-97bf-a0c41cf27c64","Type":"ContainerStarted","Data":"219cfd56b75c83080edcb3b6544f01c797143f33d364254085fd7e887a069698"} Dec 04 19:25:14 crc kubenswrapper[4733]: I1204 19:25:14.665133 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-55gp7"] Dec 04 19:25:14 crc kubenswrapper[4733]: I1204 19:25:14.666288 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-55gp7" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" containerName="registry-server" containerID="cri-o://2102c1c79943dfe4d112d370782cf5152d46286fe6862bc1f0094a2a7ff235f5" gracePeriod=2 Dec 04 19:25:14 crc kubenswrapper[4733]: E1204 19:25:14.884225 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8e002d8_550d_4ebc_8a15_52a801602065.slice/crio-conmon-2102c1c79943dfe4d112d370782cf5152d46286fe6862bc1f0094a2a7ff235f5.scope\": RecentStats: unable to find data in memory cache]" Dec 04 19:25:15 crc kubenswrapper[4733]: I1204 19:25:15.616495 4733 generic.go:334] "Generic (PLEG): container finished" podID="a8e002d8-550d-4ebc-8a15-52a801602065" containerID="2102c1c79943dfe4d112d370782cf5152d46286fe6862bc1f0094a2a7ff235f5" exitCode=0 Dec 04 19:25:15 crc kubenswrapper[4733]: I1204 19:25:15.617139 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55gp7" event={"ID":"a8e002d8-550d-4ebc-8a15-52a801602065","Type":"ContainerDied","Data":"2102c1c79943dfe4d112d370782cf5152d46286fe6862bc1f0094a2a7ff235f5"} Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.079775 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.192087 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-utilities\") pod \"a8e002d8-550d-4ebc-8a15-52a801602065\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.192165 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjm7z\" (UniqueName: \"kubernetes.io/projected/a8e002d8-550d-4ebc-8a15-52a801602065-kube-api-access-gjm7z\") pod \"a8e002d8-550d-4ebc-8a15-52a801602065\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.192289 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-catalog-content\") pod \"a8e002d8-550d-4ebc-8a15-52a801602065\" (UID: \"a8e002d8-550d-4ebc-8a15-52a801602065\") " Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.193307 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-utilities" (OuterVolumeSpecName: "utilities") pod "a8e002d8-550d-4ebc-8a15-52a801602065" (UID: "a8e002d8-550d-4ebc-8a15-52a801602065"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.201332 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8e002d8-550d-4ebc-8a15-52a801602065-kube-api-access-gjm7z" (OuterVolumeSpecName: "kube-api-access-gjm7z") pod "a8e002d8-550d-4ebc-8a15-52a801602065" (UID: "a8e002d8-550d-4ebc-8a15-52a801602065"). InnerVolumeSpecName "kube-api-access-gjm7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.228762 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8e002d8-550d-4ebc-8a15-52a801602065" (UID: "a8e002d8-550d-4ebc-8a15-52a801602065"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.295418 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.295449 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjm7z\" (UniqueName: \"kubernetes.io/projected/a8e002d8-550d-4ebc-8a15-52a801602065-kube-api-access-gjm7z\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.295460 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e002d8-550d-4ebc-8a15-52a801602065-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.630442 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55gp7" event={"ID":"a8e002d8-550d-4ebc-8a15-52a801602065","Type":"ContainerDied","Data":"5274700e19e54b4f7980fda6988695097975e733a3cef88107e09fa04a6bd581"} Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.630752 4733 scope.go:117] "RemoveContainer" containerID="2102c1c79943dfe4d112d370782cf5152d46286fe6862bc1f0094a2a7ff235f5" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.630470 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-55gp7" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.633511 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" event={"ID":"37deee19-62e0-46de-bf27-99748ce7a98b","Type":"ContainerStarted","Data":"6ffffcd289106f36c8fdbc1820362f4c31249b155316a6a6efb18fd604db34d9"} Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.637569 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" event={"ID":"7f54ccac-c4f7-47fe-b071-87d8c3231848","Type":"ContainerStarted","Data":"af41af7711c7972c0d7895d6f19fd779ed630f8807259151c8aa6b61407f84f6"} Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.656605 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q" podStartSLOduration=2.180056653 podStartE2EDuration="5.65658436s" podCreationTimestamp="2025-12-04 19:25:11 +0000 UTC" firstStartedPulling="2025-12-04 19:25:12.801210893 +0000 UTC m=+6374.756571939" lastFinishedPulling="2025-12-04 19:25:16.2777386 +0000 UTC m=+6378.233099646" observedRunningTime="2025-12-04 19:25:16.653126816 +0000 UTC m=+6378.608487862" watchObservedRunningTime="2025-12-04 19:25:16.65658436 +0000 UTC m=+6378.611945406" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.704437 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-55gp7"] Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.721651 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-55gp7"] Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.724953 4733 scope.go:117] "RemoveContainer" containerID="eb9642d96d281fe97053ec79698b610d15dcd29c3f5ac52bca00a43a0550aeab" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.755521 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp" podStartSLOduration=2.911148381 podStartE2EDuration="5.755503028s" podCreationTimestamp="2025-12-04 19:25:11 +0000 UTC" firstStartedPulling="2025-12-04 19:25:12.787325067 +0000 UTC m=+6374.742686113" lastFinishedPulling="2025-12-04 19:25:15.631679714 +0000 UTC m=+6377.587040760" observedRunningTime="2025-12-04 19:25:16.724999072 +0000 UTC m=+6378.680360118" watchObservedRunningTime="2025-12-04 19:25:16.755503028 +0000 UTC m=+6378.710864074" Dec 04 19:25:16 crc kubenswrapper[4733]: I1204 19:25:16.773862 4733 scope.go:117] "RemoveContainer" containerID="93f26c35ee80266acc3eaeaf3ed182b6c15dda22a289cad403566324f2527d42" Dec 04 19:25:18 crc kubenswrapper[4733]: I1204 19:25:18.353835 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" path="/var/lib/kubelet/pods/a8e002d8-550d-4ebc-8a15-52a801602065/volumes" Dec 04 19:25:20 crc kubenswrapper[4733]: I1204 19:25:20.699673 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-95nbh" event={"ID":"3e69f6b5-cf40-4ebe-9949-71031eb10fc8","Type":"ContainerStarted","Data":"6541db854d410d44a8fabb0338a99e1b65697dee3f19b352af9022893cb749ec"} Dec 04 19:25:20 crc kubenswrapper[4733]: I1204 19:25:20.700133 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:20 crc kubenswrapper[4733]: I1204 19:25:20.702054 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r" event={"ID":"1aa893e4-e63a-4f24-97bf-a0c41cf27c64","Type":"ContainerStarted","Data":"8993da76eadb6870f8bc00e830d8b2125ad435ccb25cb7df3c5c93f93471ed85"} Dec 04 19:25:20 crc kubenswrapper[4733]: I1204 19:25:20.751855 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-95nbh" podStartSLOduration=3.165414507 podStartE2EDuration="9.751830153s" podCreationTimestamp="2025-12-04 19:25:11 +0000 UTC" firstStartedPulling="2025-12-04 19:25:12.944917234 +0000 UTC m=+6374.900278290" lastFinishedPulling="2025-12-04 19:25:19.53133289 +0000 UTC m=+6381.486693936" observedRunningTime="2025-12-04 19:25:20.74510576 +0000 UTC m=+6382.700466816" watchObservedRunningTime="2025-12-04 19:25:20.751830153 +0000 UTC m=+6382.707191219" Dec 04 19:25:20 crc kubenswrapper[4733]: I1204 19:25:20.793767 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5lv7r" podStartSLOduration=3.021134971 podStartE2EDuration="9.793737658s" podCreationTimestamp="2025-12-04 19:25:11 +0000 UTC" firstStartedPulling="2025-12-04 19:25:12.787706717 +0000 UTC m=+6374.743067763" lastFinishedPulling="2025-12-04 19:25:19.560309404 +0000 UTC m=+6381.515670450" observedRunningTime="2025-12-04 19:25:20.767832845 +0000 UTC m=+6382.723193901" watchObservedRunningTime="2025-12-04 19:25:20.793737658 +0000 UTC m=+6382.749098704" Dec 04 19:25:24 crc kubenswrapper[4733]: I1204 19:25:24.052212 4733 scope.go:117] "RemoveContainer" containerID="b42d4642c74b57585a1c8002ad88a3d3181f2a5e79c9190632788fb62733d585" Dec 04 19:25:24 crc kubenswrapper[4733]: I1204 19:25:24.698518 4733 scope.go:117] "RemoveContainer" containerID="f6dbdc315efef701e580c60f2c1db14ec2b00a9d7b4bd2fe8bdfb0533a5091fe" Dec 04 19:25:24 crc kubenswrapper[4733]: I1204 19:25:24.748470 4733 scope.go:117] "RemoveContainer" containerID="f3459b98e6bd5c66f51153a0d7f2b5801385e9875f57698c03c3a40b15965f6d" Dec 04 19:25:26 crc kubenswrapper[4733]: I1204 19:25:26.762730 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" event={"ID":"97bb703c-e296-4b6d-b878-659c7149d970","Type":"ContainerStarted","Data":"39374cf368c694ba7833a99ba421b892bbb343884ff5657c656146c66dab000b"} Dec 04 19:25:26 crc kubenswrapper[4733]: I1204 19:25:26.763410 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:26 crc kubenswrapper[4733]: I1204 19:25:26.766201 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" Dec 04 19:25:26 crc kubenswrapper[4733]: I1204 19:25:26.790928 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-pjs6d" podStartSLOduration=2.758169537 podStartE2EDuration="15.790904785s" podCreationTimestamp="2025-12-04 19:25:11 +0000 UTC" firstStartedPulling="2025-12-04 19:25:12.843848487 +0000 UTC m=+6374.799209533" lastFinishedPulling="2025-12-04 19:25:25.876583745 +0000 UTC m=+6387.831944781" observedRunningTime="2025-12-04 19:25:26.782848508 +0000 UTC m=+6388.738209554" watchObservedRunningTime="2025-12-04 19:25:26.790904785 +0000 UTC m=+6388.746265831" Dec 04 19:25:32 crc kubenswrapper[4733]: I1204 19:25:32.076191 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-95nbh" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.503135 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.503849 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" containerName="openstackclient" containerID="cri-o://28e79a2f25597ff7c04d43e40505344eadc98163b4e687c090db85ca610569c9" gracePeriod=2 Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.519741 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.583637 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 19:25:34 crc kubenswrapper[4733]: E1204 19:25:34.584050 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" containerName="openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.584067 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" containerName="openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: E1204 19:25:34.584093 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" containerName="extract-utilities" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.584101 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" containerName="extract-utilities" Dec 04 19:25:34 crc kubenswrapper[4733]: E1204 19:25:34.584111 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" containerName="extract-content" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.584119 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" containerName="extract-content" Dec 04 19:25:34 crc kubenswrapper[4733]: E1204 19:25:34.584144 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" containerName="registry-server" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.584149 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" containerName="registry-server" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.584337 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8e002d8-550d-4ebc-8a15-52a801602065" containerName="registry-server" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.584347 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" containerName="openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.587101 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.595187 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.604530 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" podUID="4dcf8409-c5dd-4cf4-bffd-c08a965fca43" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.665037 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-openstack-config-secret\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.665095 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-openstack-config\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.665136 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdds9\" (UniqueName: \"kubernetes.io/projected/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-kube-api-access-hdds9\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.767264 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-openstack-config-secret\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.767350 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-openstack-config\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.767394 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdds9\" (UniqueName: \"kubernetes.io/projected/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-kube-api-access-hdds9\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.768760 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-openstack-config\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.775215 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-openstack-config-secret\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.802455 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdds9\" (UniqueName: \"kubernetes.io/projected/4dcf8409-c5dd-4cf4-bffd-c08a965fca43-kube-api-access-hdds9\") pod \"openstackclient\" (UID: \"4dcf8409-c5dd-4cf4-bffd-c08a965fca43\") " pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.850223 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.865335 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.868805 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-hhjst" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.870322 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.914328 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:25:34 crc kubenswrapper[4733]: I1204 19:25:34.982320 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf6fr\" (UniqueName: \"kubernetes.io/projected/a2fce55f-036a-419c-b710-17eafaa65162-kube-api-access-mf6fr\") pod \"kube-state-metrics-0\" (UID: \"a2fce55f-036a-419c-b710-17eafaa65162\") " pod="openstack/kube-state-metrics-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.089118 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf6fr\" (UniqueName: \"kubernetes.io/projected/a2fce55f-036a-419c-b710-17eafaa65162-kube-api-access-mf6fr\") pod \"kube-state-metrics-0\" (UID: \"a2fce55f-036a-419c-b710-17eafaa65162\") " pod="openstack/kube-state-metrics-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.128823 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf6fr\" (UniqueName: \"kubernetes.io/projected/a2fce55f-036a-419c-b710-17eafaa65162-kube-api-access-mf6fr\") pod \"kube-state-metrics-0\" (UID: \"a2fce55f-036a-419c-b710-17eafaa65162\") " pod="openstack/kube-state-metrics-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.226673 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.647266 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.650176 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.656316 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.656393 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.656542 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.656651 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.656751 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-z9jr5" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.677365 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.713228 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4ml9\" (UniqueName: \"kubernetes.io/projected/308b5993-05c7-4915-b744-8ac8797da024-kube-api-access-r4ml9\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.713291 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.713368 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/308b5993-05c7-4915-b744-8ac8797da024-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.713391 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/308b5993-05c7-4915-b744-8ac8797da024-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.713418 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/308b5993-05c7-4915-b744-8ac8797da024-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.713462 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.713480 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.814716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/308b5993-05c7-4915-b744-8ac8797da024-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.815045 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/308b5993-05c7-4915-b744-8ac8797da024-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.815069 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/308b5993-05c7-4915-b744-8ac8797da024-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.815111 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.815130 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.815190 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4ml9\" (UniqueName: \"kubernetes.io/projected/308b5993-05c7-4915-b744-8ac8797da024-kube-api-access-r4ml9\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.815217 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.816595 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/308b5993-05c7-4915-b744-8ac8797da024-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.824640 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.825697 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/308b5993-05c7-4915-b744-8ac8797da024-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.827260 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.829364 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/308b5993-05c7-4915-b744-8ac8797da024-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.842158 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/308b5993-05c7-4915-b744-8ac8797da024-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:35 crc kubenswrapper[4733]: I1204 19:25:35.878015 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4ml9\" (UniqueName: \"kubernetes.io/projected/308b5993-05c7-4915-b744-8ac8797da024-kube-api-access-r4ml9\") pod \"alertmanager-metric-storage-0\" (UID: \"308b5993-05c7-4915-b744-8ac8797da024\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.039775 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.161615 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.175008 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.178591 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.181394 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.181780 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.196166 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.196403 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.197857 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-lpbwc" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.259108 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.259192 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-config\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.259245 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.259339 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.259460 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7cb310db-5ba0-4b88-9cef-0efa33e631e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7cb310db-5ba0-4b88-9cef-0efa33e631e1\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.259516 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpbtb\" (UniqueName: \"kubernetes.io/projected/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-kube-api-access-vpbtb\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.259654 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.259710 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.385916 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.386009 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.386059 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.386087 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-config\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.386121 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.386168 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.386226 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7cb310db-5ba0-4b88-9cef-0efa33e631e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7cb310db-5ba0-4b88-9cef-0efa33e631e1\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.393665 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.405681 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.406037 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-config\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.415783 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpbtb\" (UniqueName: \"kubernetes.io/projected/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-kube-api-access-vpbtb\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.436725 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.436784 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.443342 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.448118 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.455250 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.479321 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpbtb\" (UniqueName: \"kubernetes.io/projected/aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf-kube-api-access-vpbtb\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.498397 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.498440 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7cb310db-5ba0-4b88-9cef-0efa33e631e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7cb310db-5ba0-4b88-9cef-0efa33e631e1\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/071f9e85d8d3c7ee1eb0a360bc8196c619269abc0eaa085b0903a0fea5ab4838/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.711643 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.749579 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7cb310db-5ba0-4b88-9cef-0efa33e631e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7cb310db-5ba0-4b88-9cef-0efa33e631e1\") pod \"prometheus-metric-storage-0\" (UID: \"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf\") " pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.786430 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.898529 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4dcf8409-c5dd-4cf4-bffd-c08a965fca43","Type":"ContainerStarted","Data":"6ffa0974e29c7b6ba52ffa7ce353a5c751eae52ed1819532ad7f0b13719eb7fe"} Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.899843 4733 generic.go:334] "Generic (PLEG): container finished" podID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" containerID="28e79a2f25597ff7c04d43e40505344eadc98163b4e687c090db85ca610569c9" exitCode=137 Dec 04 19:25:36 crc kubenswrapper[4733]: I1204 19:25:36.904322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a2fce55f-036a-419c-b710-17eafaa65162","Type":"ContainerStarted","Data":"096f5135ac805a4b6d1251bdbe36f0017be7259caecec82f51c66de568769ce0"} Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.177209 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.468180 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 19:25:37 crc kubenswrapper[4733]: W1204 19:25:37.492952 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaede4a3c_0a99_45a2_a1cc_833a9f8cb6bf.slice/crio-a761f46b367e413a5e2131f56bfd400afc805305f05fd2f5edb6d78f59c37ef2 WatchSource:0}: Error finding container a761f46b367e413a5e2131f56bfd400afc805305f05fd2f5edb6d78f59c37ef2: Status 404 returned error can't find the container with id a761f46b367e413a5e2131f56bfd400afc805305f05fd2f5edb6d78f59c37ef2 Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.497818 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.557785 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmrs4\" (UniqueName: \"kubernetes.io/projected/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-kube-api-access-pmrs4\") pod \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.558200 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config\") pod \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.558396 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config-secret\") pod \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\" (UID: \"9216b0d2-71e0-4fe9-9d6d-3444e1d795b5\") " Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.572756 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-kube-api-access-pmrs4" (OuterVolumeSpecName: "kube-api-access-pmrs4") pod "9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" (UID: "9216b0d2-71e0-4fe9-9d6d-3444e1d795b5"). InnerVolumeSpecName "kube-api-access-pmrs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.605342 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" (UID: "9216b0d2-71e0-4fe9-9d6d-3444e1d795b5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.623146 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" (UID: "9216b0d2-71e0-4fe9-9d6d-3444e1d795b5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.660201 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmrs4\" (UniqueName: \"kubernetes.io/projected/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-kube-api-access-pmrs4\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.660235 4733 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.660244 4733 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.913245 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf","Type":"ContainerStarted","Data":"a761f46b367e413a5e2131f56bfd400afc805305f05fd2f5edb6d78f59c37ef2"} Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.915439 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a2fce55f-036a-419c-b710-17eafaa65162","Type":"ContainerStarted","Data":"fd8a4163471cf4dccaeec61a292b5a338a6163e40bf0c1d6b311404a7604c26d"} Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.915629 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.917234 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4dcf8409-c5dd-4cf4-bffd-c08a965fca43","Type":"ContainerStarted","Data":"22c977a10bd76aa56c81af20491e7c7db8e9efb8473dda9ddc1141c1ec04a8d6"} Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.919229 4733 scope.go:117] "RemoveContainer" containerID="28e79a2f25597ff7c04d43e40505344eadc98163b4e687c090db85ca610569c9" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.919372 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.920994 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"308b5993-05c7-4915-b744-8ac8797da024","Type":"ContainerStarted","Data":"f4127fead097a5f97acc5d92f2f840556c069e405253791dc0794739ce38f631"} Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.941845 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.367378777 podStartE2EDuration="3.941821243s" podCreationTimestamp="2025-12-04 19:25:34 +0000 UTC" firstStartedPulling="2025-12-04 19:25:36.818225725 +0000 UTC m=+6398.773586771" lastFinishedPulling="2025-12-04 19:25:37.392668191 +0000 UTC m=+6399.348029237" observedRunningTime="2025-12-04 19:25:37.932699636 +0000 UTC m=+6399.888060682" watchObservedRunningTime="2025-12-04 19:25:37.941821243 +0000 UTC m=+6399.897182289" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.962095 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" podUID="4dcf8409-c5dd-4cf4-bffd-c08a965fca43" Dec 04 19:25:37 crc kubenswrapper[4733]: I1204 19:25:37.963865 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.963773857 podStartE2EDuration="3.963773857s" podCreationTimestamp="2025-12-04 19:25:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:25:37.956974843 +0000 UTC m=+6399.912335889" watchObservedRunningTime="2025-12-04 19:25:37.963773857 +0000 UTC m=+6399.919134903" Dec 04 19:25:38 crc kubenswrapper[4733]: I1204 19:25:38.360269 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9216b0d2-71e0-4fe9-9d6d-3444e1d795b5" path="/var/lib/kubelet/pods/9216b0d2-71e0-4fe9-9d6d-3444e1d795b5/volumes" Dec 04 19:25:39 crc kubenswrapper[4733]: I1204 19:25:39.200334 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-hbrmg"] Dec 04 19:25:39 crc kubenswrapper[4733]: I1204 19:25:39.206919 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-41e5-account-create-update-w82s7"] Dec 04 19:25:39 crc kubenswrapper[4733]: I1204 19:25:39.230306 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-hbrmg"] Dec 04 19:25:39 crc kubenswrapper[4733]: I1204 19:25:39.284224 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-41e5-account-create-update-w82s7"] Dec 04 19:25:40 crc kubenswrapper[4733]: I1204 19:25:40.345775 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dfb8e08-b816-4747-a6a0-a0e72ff9acec" path="/var/lib/kubelet/pods/8dfb8e08-b816-4747-a6a0-a0e72ff9acec/volumes" Dec 04 19:25:40 crc kubenswrapper[4733]: I1204 19:25:40.347125 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a4146c6-9c30-4565-b821-3bd9c60e9917" path="/var/lib/kubelet/pods/9a4146c6-9c30-4565-b821-3bd9c60e9917/volumes" Dec 04 19:25:41 crc kubenswrapper[4733]: I1204 19:25:41.065184 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8c5f-account-create-update-m2vkx"] Dec 04 19:25:41 crc kubenswrapper[4733]: I1204 19:25:41.083557 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-633b-account-create-update-wx467"] Dec 04 19:25:41 crc kubenswrapper[4733]: I1204 19:25:41.098242 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-rv6rz"] Dec 04 19:25:41 crc kubenswrapper[4733]: I1204 19:25:41.108271 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-h29sl"] Dec 04 19:25:41 crc kubenswrapper[4733]: I1204 19:25:41.118554 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-h29sl"] Dec 04 19:25:41 crc kubenswrapper[4733]: I1204 19:25:41.130962 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-rv6rz"] Dec 04 19:25:41 crc kubenswrapper[4733]: I1204 19:25:41.138460 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8c5f-account-create-update-m2vkx"] Dec 04 19:25:41 crc kubenswrapper[4733]: I1204 19:25:41.147763 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-633b-account-create-update-wx467"] Dec 04 19:25:42 crc kubenswrapper[4733]: I1204 19:25:42.363353 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="628cfd34-aecf-4401-961d-42e8b4884a5d" path="/var/lib/kubelet/pods/628cfd34-aecf-4401-961d-42e8b4884a5d/volumes" Dec 04 19:25:42 crc kubenswrapper[4733]: I1204 19:25:42.364454 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="673f8f69-9bd9-40c3-b67b-26f43f04afe7" path="/var/lib/kubelet/pods/673f8f69-9bd9-40c3-b67b-26f43f04afe7/volumes" Dec 04 19:25:42 crc kubenswrapper[4733]: I1204 19:25:42.365137 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94ac3286-fd11-4ffa-9da1-c35564d4e7a6" path="/var/lib/kubelet/pods/94ac3286-fd11-4ffa-9da1-c35564d4e7a6/volumes" Dec 04 19:25:42 crc kubenswrapper[4733]: I1204 19:25:42.365852 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a7bee5-b6c4-4f00-8317-af0fe4133000" path="/var/lib/kubelet/pods/97a7bee5-b6c4-4f00-8317-af0fe4133000/volumes" Dec 04 19:25:45 crc kubenswrapper[4733]: I1204 19:25:45.018176 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf","Type":"ContainerStarted","Data":"5007af487b47f0241a012ec2a62eb4422ebc1f05aeacc19c415a94f94e64de64"} Dec 04 19:25:45 crc kubenswrapper[4733]: I1204 19:25:45.033549 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"308b5993-05c7-4915-b744-8ac8797da024","Type":"ContainerStarted","Data":"7811f333aea493bdf09ffaee52fad656660a602e0af6acf6f0b1132bcadf4533"} Dec 04 19:25:45 crc kubenswrapper[4733]: I1204 19:25:45.231487 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 19:25:45 crc kubenswrapper[4733]: I1204 19:25:45.361764 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:25:45 crc kubenswrapper[4733]: I1204 19:25:45.361834 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:25:51 crc kubenswrapper[4733]: I1204 19:25:51.132913 4733 generic.go:334] "Generic (PLEG): container finished" podID="aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf" containerID="5007af487b47f0241a012ec2a62eb4422ebc1f05aeacc19c415a94f94e64de64" exitCode=0 Dec 04 19:25:51 crc kubenswrapper[4733]: I1204 19:25:51.133046 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf","Type":"ContainerDied","Data":"5007af487b47f0241a012ec2a62eb4422ebc1f05aeacc19c415a94f94e64de64"} Dec 04 19:25:52 crc kubenswrapper[4733]: I1204 19:25:52.145218 4733 generic.go:334] "Generic (PLEG): container finished" podID="308b5993-05c7-4915-b744-8ac8797da024" containerID="7811f333aea493bdf09ffaee52fad656660a602e0af6acf6f0b1132bcadf4533" exitCode=0 Dec 04 19:25:52 crc kubenswrapper[4733]: I1204 19:25:52.145270 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"308b5993-05c7-4915-b744-8ac8797da024","Type":"ContainerDied","Data":"7811f333aea493bdf09ffaee52fad656660a602e0af6acf6f0b1132bcadf4533"} Dec 04 19:25:54 crc kubenswrapper[4733]: I1204 19:25:54.038963 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-shzgf"] Dec 04 19:25:54 crc kubenswrapper[4733]: I1204 19:25:54.051640 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-shzgf"] Dec 04 19:25:54 crc kubenswrapper[4733]: I1204 19:25:54.354090 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82b804be-19c0-4315-9104-7c4edba4cb66" path="/var/lib/kubelet/pods/82b804be-19c0-4315-9104-7c4edba4cb66/volumes" Dec 04 19:25:56 crc kubenswrapper[4733]: I1204 19:25:56.201188 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"308b5993-05c7-4915-b744-8ac8797da024","Type":"ContainerStarted","Data":"1cdf1703418030e908d9ae9b622d7afe98463a5235e75d2afe15c548730f810f"} Dec 04 19:25:56 crc kubenswrapper[4733]: I1204 19:25:56.206216 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf","Type":"ContainerStarted","Data":"7800210307f2455d46ce0311b9bfc0f0d47c21eeb5ec0dfb6b8d3f0a05bd0c66"} Dec 04 19:26:02 crc kubenswrapper[4733]: I1204 19:26:02.278440 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf","Type":"ContainerStarted","Data":"9d511dcfdebe8df5d992fc56d5dda09852d6c05dba1dd798ccbf6ecef5b079c5"} Dec 04 19:26:02 crc kubenswrapper[4733]: I1204 19:26:02.281675 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"308b5993-05c7-4915-b744-8ac8797da024","Type":"ContainerStarted","Data":"b4f21a76e0410c085b275e36dce290403559867ab583bc71227f774a27a98cdf"} Dec 04 19:26:02 crc kubenswrapper[4733]: I1204 19:26:02.324754 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=8.65935969 podStartE2EDuration="27.324733095s" podCreationTimestamp="2025-12-04 19:25:35 +0000 UTC" firstStartedPulling="2025-12-04 19:25:37.197348822 +0000 UTC m=+6399.152709868" lastFinishedPulling="2025-12-04 19:25:55.862722227 +0000 UTC m=+6417.818083273" observedRunningTime="2025-12-04 19:26:02.317523999 +0000 UTC m=+6424.272885125" watchObservedRunningTime="2025-12-04 19:26:02.324733095 +0000 UTC m=+6424.280094141" Dec 04 19:26:03 crc kubenswrapper[4733]: I1204 19:26:03.296408 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 04 19:26:03 crc kubenswrapper[4733]: I1204 19:26:03.300959 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 04 19:26:07 crc kubenswrapper[4733]: I1204 19:26:07.040133 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z4fwn"] Dec 04 19:26:07 crc kubenswrapper[4733]: I1204 19:26:07.055783 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z4fwn"] Dec 04 19:26:07 crc kubenswrapper[4733]: I1204 19:26:07.344753 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf","Type":"ContainerStarted","Data":"a28536dcbafb7787ff9fecae13d6127e11bf1ced4f7d5dd12c135724bbc3a3f8"} Dec 04 19:26:07 crc kubenswrapper[4733]: I1204 19:26:07.371785 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.451986218 podStartE2EDuration="32.371766902s" podCreationTimestamp="2025-12-04 19:25:35 +0000 UTC" firstStartedPulling="2025-12-04 19:25:37.498088857 +0000 UTC m=+6399.453449903" lastFinishedPulling="2025-12-04 19:26:06.417869531 +0000 UTC m=+6428.373230587" observedRunningTime="2025-12-04 19:26:07.366757318 +0000 UTC m=+6429.322118374" watchObservedRunningTime="2025-12-04 19:26:07.371766902 +0000 UTC m=+6429.327127948" Dec 04 19:26:08 crc kubenswrapper[4733]: I1204 19:26:08.038520 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-5j49q"] Dec 04 19:26:08 crc kubenswrapper[4733]: I1204 19:26:08.050151 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-5j49q"] Dec 04 19:26:08 crc kubenswrapper[4733]: I1204 19:26:08.368525 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bcff476-b8d5-4ddd-91e6-25e5c7488b57" path="/var/lib/kubelet/pods/3bcff476-b8d5-4ddd-91e6-25e5c7488b57/volumes" Dec 04 19:26:08 crc kubenswrapper[4733]: I1204 19:26:08.370290 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c44dddb-58bb-476e-91d3-5d0663897c2e" path="/var/lib/kubelet/pods/7c44dddb-58bb-476e-91d3-5d0663897c2e/volumes" Dec 04 19:26:11 crc kubenswrapper[4733]: I1204 19:26:11.788410 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.864851 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.867826 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.870688 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.871074 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.883839 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.961702 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-config-data\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.961957 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.962033 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.962117 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-scripts\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.962189 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6brld\" (UniqueName: \"kubernetes.io/projected/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-kube-api-access-6brld\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.962267 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-run-httpd\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:14 crc kubenswrapper[4733]: I1204 19:26:14.962348 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-log-httpd\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.063544 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-run-httpd\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.064187 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-log-httpd\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.064607 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-config-data\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.065329 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.065428 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.065536 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-scripts\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.065627 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6brld\" (UniqueName: \"kubernetes.io/projected/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-kube-api-access-6brld\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.064462 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-log-httpd\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.064137 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-run-httpd\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.083437 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.086434 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.087707 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-scripts\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.087714 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-config-data\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.112781 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6brld\" (UniqueName: \"kubernetes.io/projected/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-kube-api-access-6brld\") pod \"ceilometer-0\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.189577 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.361862 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.362194 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:26:15 crc kubenswrapper[4733]: I1204 19:26:15.623472 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:15 crc kubenswrapper[4733]: W1204 19:26:15.624420 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a3cbb1c_22cd_4e94_8e52_b9f63a5d13cf.slice/crio-5faca64471e0b151a655805eed16f7387fab63c35e28b3c5118c3b362f6c381b WatchSource:0}: Error finding container 5faca64471e0b151a655805eed16f7387fab63c35e28b3c5118c3b362f6c381b: Status 404 returned error can't find the container with id 5faca64471e0b151a655805eed16f7387fab63c35e28b3c5118c3b362f6c381b Dec 04 19:26:16 crc kubenswrapper[4733]: I1204 19:26:16.450495 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerStarted","Data":"5faca64471e0b151a655805eed16f7387fab63c35e28b3c5118c3b362f6c381b"} Dec 04 19:26:17 crc kubenswrapper[4733]: I1204 19:26:17.463052 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerStarted","Data":"961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f"} Dec 04 19:26:17 crc kubenswrapper[4733]: I1204 19:26:17.463675 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerStarted","Data":"487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f"} Dec 04 19:26:18 crc kubenswrapper[4733]: I1204 19:26:18.474361 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerStarted","Data":"341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5"} Dec 04 19:26:19 crc kubenswrapper[4733]: I1204 19:26:19.488630 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerStarted","Data":"fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb"} Dec 04 19:26:19 crc kubenswrapper[4733]: I1204 19:26:19.489166 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 19:26:19 crc kubenswrapper[4733]: I1204 19:26:19.519266 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.025064862 podStartE2EDuration="5.519243038s" podCreationTimestamp="2025-12-04 19:26:14 +0000 UTC" firstStartedPulling="2025-12-04 19:26:15.628150273 +0000 UTC m=+6437.583511319" lastFinishedPulling="2025-12-04 19:26:19.122328449 +0000 UTC m=+6441.077689495" observedRunningTime="2025-12-04 19:26:19.506597065 +0000 UTC m=+6441.461958121" watchObservedRunningTime="2025-12-04 19:26:19.519243038 +0000 UTC m=+6441.474604094" Dec 04 19:26:21 crc kubenswrapper[4733]: I1204 19:26:21.788046 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 04 19:26:21 crc kubenswrapper[4733]: I1204 19:26:21.791258 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 04 19:26:22 crc kubenswrapper[4733]: I1204 19:26:22.041759 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-6v6gj"] Dec 04 19:26:22 crc kubenswrapper[4733]: I1204 19:26:22.049917 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-6v6gj"] Dec 04 19:26:22 crc kubenswrapper[4733]: I1204 19:26:22.368744 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9d10c32-2f7a-4306-82cc-f84d893f2045" path="/var/lib/kubelet/pods/b9d10c32-2f7a-4306-82cc-f84d893f2045/volumes" Dec 04 19:26:22 crc kubenswrapper[4733]: I1204 19:26:22.514908 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.038650 4733 scope.go:117] "RemoveContainer" containerID="ec18881b2092dca99136a1c7fcb83ae0cadac1d0d6aa515c5abd7c66428824fa" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.076677 4733 scope.go:117] "RemoveContainer" containerID="bbab5feb0fd684ae57feacfea6fd3f334d1b1e5d5e8e136762c0eaff06c00982" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.125323 4733 scope.go:117] "RemoveContainer" containerID="9a5e9e20a72da4042f0daaffa415ae764aa94ca62de0c15aa94ac58caa04baea" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.172031 4733 scope.go:117] "RemoveContainer" containerID="d614fbaaf878ded8429b1ba8c8a7f3c1f24d8c0853b196bf494327fb43e06c25" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.212212 4733 scope.go:117] "RemoveContainer" containerID="d8204971181bedeff30bcd0c95edeff010107aac8d52df9b92c4cbcb18c54b1e" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.247050 4733 scope.go:117] "RemoveContainer" containerID="7cd08c488d0c6ca5def3ec8d745da047167f4633748fbd19c5d90476d7d7a815" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.290186 4733 scope.go:117] "RemoveContainer" containerID="1872a59391720f11d26825f0507162694ab176452ece560cbe421b193d86366e" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.309248 4733 scope.go:117] "RemoveContainer" containerID="7d3370fc0d1062dbde2b3e7e7b8b8545f725b4e2e0e812d5213d7020756f61b9" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.334892 4733 scope.go:117] "RemoveContainer" containerID="e0332ba56c48fe20015bc1c0ab7072b5413abed8bb199432361d772645e5bbf4" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.353766 4733 scope.go:117] "RemoveContainer" containerID="e0bd19e089dde943d8b5f5cdbfbddb6f3d72e7dbbde25d33e22c346bf2448024" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.382547 4733 scope.go:117] "RemoveContainer" containerID="04ef518c51ed8ba09fb4f5b2501fb3d3e0cf32c1b5a07068a443f08ff14ad56c" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.779503 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-5bf5z"] Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.781168 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.805771 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-5bf5z"] Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.937169 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcc29a4f-6e38-49ec-a8be-4464c076d391-operator-scripts\") pod \"aodh-db-create-5bf5z\" (UID: \"dcc29a4f-6e38-49ec-a8be-4464c076d391\") " pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.937538 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfhm2\" (UniqueName: \"kubernetes.io/projected/dcc29a4f-6e38-49ec-a8be-4464c076d391-kube-api-access-lfhm2\") pod \"aodh-db-create-5bf5z\" (UID: \"dcc29a4f-6e38-49ec-a8be-4464c076d391\") " pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.983725 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-af12-account-create-update-tk5q7"] Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.986424 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:25 crc kubenswrapper[4733]: I1204 19:26:25.991273 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.010082 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-af12-account-create-update-tk5q7"] Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.039533 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfhm2\" (UniqueName: \"kubernetes.io/projected/dcc29a4f-6e38-49ec-a8be-4464c076d391-kube-api-access-lfhm2\") pod \"aodh-db-create-5bf5z\" (UID: \"dcc29a4f-6e38-49ec-a8be-4464c076d391\") " pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.039639 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcc29a4f-6e38-49ec-a8be-4464c076d391-operator-scripts\") pod \"aodh-db-create-5bf5z\" (UID: \"dcc29a4f-6e38-49ec-a8be-4464c076d391\") " pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.040580 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcc29a4f-6e38-49ec-a8be-4464c076d391-operator-scripts\") pod \"aodh-db-create-5bf5z\" (UID: \"dcc29a4f-6e38-49ec-a8be-4464c076d391\") " pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.077225 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfhm2\" (UniqueName: \"kubernetes.io/projected/dcc29a4f-6e38-49ec-a8be-4464c076d391-kube-api-access-lfhm2\") pod \"aodh-db-create-5bf5z\" (UID: \"dcc29a4f-6e38-49ec-a8be-4464c076d391\") " pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.098101 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.141631 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkczd\" (UniqueName: \"kubernetes.io/projected/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-kube-api-access-fkczd\") pod \"aodh-af12-account-create-update-tk5q7\" (UID: \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\") " pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.141744 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-operator-scripts\") pod \"aodh-af12-account-create-update-tk5q7\" (UID: \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\") " pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.243124 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkczd\" (UniqueName: \"kubernetes.io/projected/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-kube-api-access-fkczd\") pod \"aodh-af12-account-create-update-tk5q7\" (UID: \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\") " pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.243406 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-operator-scripts\") pod \"aodh-af12-account-create-update-tk5q7\" (UID: \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\") " pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.244084 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-operator-scripts\") pod \"aodh-af12-account-create-update-tk5q7\" (UID: \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\") " pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.284129 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkczd\" (UniqueName: \"kubernetes.io/projected/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-kube-api-access-fkczd\") pod \"aodh-af12-account-create-update-tk5q7\" (UID: \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\") " pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.310543 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.616858 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-5bf5z"] Dec 04 19:26:26 crc kubenswrapper[4733]: I1204 19:26:26.800879 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-af12-account-create-update-tk5q7"] Dec 04 19:26:26 crc kubenswrapper[4733]: W1204 19:26:26.801868 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod449bb25e_8f81_4a6b_8b2a_dcca30e4f880.slice/crio-6f86f5da1658706d13b98e3d5e0eba696e4b24d4a5cb595534bd64401cdecbbe WatchSource:0}: Error finding container 6f86f5da1658706d13b98e3d5e0eba696e4b24d4a5cb595534bd64401cdecbbe: Status 404 returned error can't find the container with id 6f86f5da1658706d13b98e3d5e0eba696e4b24d4a5cb595534bd64401cdecbbe Dec 04 19:26:27 crc kubenswrapper[4733]: E1204 19:26:27.338152 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod449bb25e_8f81_4a6b_8b2a_dcca30e4f880.slice/crio-eed04ac06f54422891d78d3c542bc3ba195c6785e603bb08086b455165e8383f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod449bb25e_8f81_4a6b_8b2a_dcca30e4f880.slice/crio-conmon-eed04ac06f54422891d78d3c542bc3ba195c6785e603bb08086b455165e8383f.scope\": RecentStats: unable to find data in memory cache]" Dec 04 19:26:27 crc kubenswrapper[4733]: I1204 19:26:27.593141 4733 generic.go:334] "Generic (PLEG): container finished" podID="dcc29a4f-6e38-49ec-a8be-4464c076d391" containerID="97c793a24dcb898273e69f204c5adeac185a7c5755d3546329357a53a26c8b20" exitCode=0 Dec 04 19:26:27 crc kubenswrapper[4733]: I1204 19:26:27.593191 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-5bf5z" event={"ID":"dcc29a4f-6e38-49ec-a8be-4464c076d391","Type":"ContainerDied","Data":"97c793a24dcb898273e69f204c5adeac185a7c5755d3546329357a53a26c8b20"} Dec 04 19:26:27 crc kubenswrapper[4733]: I1204 19:26:27.593253 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-5bf5z" event={"ID":"dcc29a4f-6e38-49ec-a8be-4464c076d391","Type":"ContainerStarted","Data":"3dada433022ee0eeb71cbb118c2a3819470175160c927999b235775bb565563e"} Dec 04 19:26:27 crc kubenswrapper[4733]: I1204 19:26:27.596461 4733 generic.go:334] "Generic (PLEG): container finished" podID="449bb25e-8f81-4a6b-8b2a-dcca30e4f880" containerID="eed04ac06f54422891d78d3c542bc3ba195c6785e603bb08086b455165e8383f" exitCode=0 Dec 04 19:26:27 crc kubenswrapper[4733]: I1204 19:26:27.596496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-af12-account-create-update-tk5q7" event={"ID":"449bb25e-8f81-4a6b-8b2a-dcca30e4f880","Type":"ContainerDied","Data":"eed04ac06f54422891d78d3c542bc3ba195c6785e603bb08086b455165e8383f"} Dec 04 19:26:27 crc kubenswrapper[4733]: I1204 19:26:27.596511 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-af12-account-create-update-tk5q7" event={"ID":"449bb25e-8f81-4a6b-8b2a-dcca30e4f880","Type":"ContainerStarted","Data":"6f86f5da1658706d13b98e3d5e0eba696e4b24d4a5cb595534bd64401cdecbbe"} Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.084637 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.092189 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.219915 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkczd\" (UniqueName: \"kubernetes.io/projected/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-kube-api-access-fkczd\") pod \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\" (UID: \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\") " Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.220051 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfhm2\" (UniqueName: \"kubernetes.io/projected/dcc29a4f-6e38-49ec-a8be-4464c076d391-kube-api-access-lfhm2\") pod \"dcc29a4f-6e38-49ec-a8be-4464c076d391\" (UID: \"dcc29a4f-6e38-49ec-a8be-4464c076d391\") " Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.220217 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-operator-scripts\") pod \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\" (UID: \"449bb25e-8f81-4a6b-8b2a-dcca30e4f880\") " Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.220395 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcc29a4f-6e38-49ec-a8be-4464c076d391-operator-scripts\") pod \"dcc29a4f-6e38-49ec-a8be-4464c076d391\" (UID: \"dcc29a4f-6e38-49ec-a8be-4464c076d391\") " Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.220924 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcc29a4f-6e38-49ec-a8be-4464c076d391-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dcc29a4f-6e38-49ec-a8be-4464c076d391" (UID: "dcc29a4f-6e38-49ec-a8be-4464c076d391"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.220944 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "449bb25e-8f81-4a6b-8b2a-dcca30e4f880" (UID: "449bb25e-8f81-4a6b-8b2a-dcca30e4f880"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.221425 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.221443 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcc29a4f-6e38-49ec-a8be-4464c076d391-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.226433 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-kube-api-access-fkczd" (OuterVolumeSpecName: "kube-api-access-fkczd") pod "449bb25e-8f81-4a6b-8b2a-dcca30e4f880" (UID: "449bb25e-8f81-4a6b-8b2a-dcca30e4f880"). InnerVolumeSpecName "kube-api-access-fkczd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.232687 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcc29a4f-6e38-49ec-a8be-4464c076d391-kube-api-access-lfhm2" (OuterVolumeSpecName: "kube-api-access-lfhm2") pod "dcc29a4f-6e38-49ec-a8be-4464c076d391" (UID: "dcc29a4f-6e38-49ec-a8be-4464c076d391"). InnerVolumeSpecName "kube-api-access-lfhm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.322918 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkczd\" (UniqueName: \"kubernetes.io/projected/449bb25e-8f81-4a6b-8b2a-dcca30e4f880-kube-api-access-fkczd\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.322956 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfhm2\" (UniqueName: \"kubernetes.io/projected/dcc29a4f-6e38-49ec-a8be-4464c076d391-kube-api-access-lfhm2\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.622639 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-5bf5z" event={"ID":"dcc29a4f-6e38-49ec-a8be-4464c076d391","Type":"ContainerDied","Data":"3dada433022ee0eeb71cbb118c2a3819470175160c927999b235775bb565563e"} Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.622682 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dada433022ee0eeb71cbb118c2a3819470175160c927999b235775bb565563e" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.622732 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-5bf5z" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.624160 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-af12-account-create-update-tk5q7" event={"ID":"449bb25e-8f81-4a6b-8b2a-dcca30e4f880","Type":"ContainerDied","Data":"6f86f5da1658706d13b98e3d5e0eba696e4b24d4a5cb595534bd64401cdecbbe"} Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.624182 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f86f5da1658706d13b98e3d5e0eba696e4b24d4a5cb595534bd64401cdecbbe" Dec 04 19:26:29 crc kubenswrapper[4733]: I1204 19:26:29.624415 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-af12-account-create-update-tk5q7" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.348517 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-7jggv"] Dec 04 19:26:31 crc kubenswrapper[4733]: E1204 19:26:31.350074 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc29a4f-6e38-49ec-a8be-4464c076d391" containerName="mariadb-database-create" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.350100 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc29a4f-6e38-49ec-a8be-4464c076d391" containerName="mariadb-database-create" Dec 04 19:26:31 crc kubenswrapper[4733]: E1204 19:26:31.350182 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449bb25e-8f81-4a6b-8b2a-dcca30e4f880" containerName="mariadb-account-create-update" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.350191 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="449bb25e-8f81-4a6b-8b2a-dcca30e4f880" containerName="mariadb-account-create-update" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.354077 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc29a4f-6e38-49ec-a8be-4464c076d391" containerName="mariadb-database-create" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.354151 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="449bb25e-8f81-4a6b-8b2a-dcca30e4f880" containerName="mariadb-account-create-update" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.355548 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.357850 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.360240 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.361442 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.361672 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-kc27c" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.385852 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-7jggv"] Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.467441 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzbjs\" (UniqueName: \"kubernetes.io/projected/9595436c-1e86-4790-8ca4-3707f0d22b0c-kube-api-access-qzbjs\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.467541 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-config-data\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.467586 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-scripts\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.467658 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-combined-ca-bundle\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.569465 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-combined-ca-bundle\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.569600 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzbjs\" (UniqueName: \"kubernetes.io/projected/9595436c-1e86-4790-8ca4-3707f0d22b0c-kube-api-access-qzbjs\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.569640 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-config-data\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.569673 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-scripts\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.575462 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-scripts\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.575783 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-config-data\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.577080 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-combined-ca-bundle\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.603585 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzbjs\" (UniqueName: \"kubernetes.io/projected/9595436c-1e86-4790-8ca4-3707f0d22b0c-kube-api-access-qzbjs\") pod \"aodh-db-sync-7jggv\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:31 crc kubenswrapper[4733]: I1204 19:26:31.676401 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:32 crc kubenswrapper[4733]: W1204 19:26:32.189813 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9595436c_1e86_4790_8ca4_3707f0d22b0c.slice/crio-4004a2eddc9316c32395d80d69de10bbb8726e0de0b0faac96a3dd185e3cc721 WatchSource:0}: Error finding container 4004a2eddc9316c32395d80d69de10bbb8726e0de0b0faac96a3dd185e3cc721: Status 404 returned error can't find the container with id 4004a2eddc9316c32395d80d69de10bbb8726e0de0b0faac96a3dd185e3cc721 Dec 04 19:26:32 crc kubenswrapper[4733]: I1204 19:26:32.194219 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-7jggv"] Dec 04 19:26:32 crc kubenswrapper[4733]: I1204 19:26:32.660665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-7jggv" event={"ID":"9595436c-1e86-4790-8ca4-3707f0d22b0c","Type":"ContainerStarted","Data":"4004a2eddc9316c32395d80d69de10bbb8726e0de0b0faac96a3dd185e3cc721"} Dec 04 19:26:37 crc kubenswrapper[4733]: I1204 19:26:37.727079 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-7jggv" event={"ID":"9595436c-1e86-4790-8ca4-3707f0d22b0c","Type":"ContainerStarted","Data":"7d37f601d15b0a66fc6301a369fbab32c1f52dfce3a3eb81b04d53cb4228476e"} Dec 04 19:26:37 crc kubenswrapper[4733]: I1204 19:26:37.752165 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-7jggv" podStartSLOduration=2.404940335 podStartE2EDuration="6.752147801s" podCreationTimestamp="2025-12-04 19:26:31 +0000 UTC" firstStartedPulling="2025-12-04 19:26:32.191839813 +0000 UTC m=+6454.147200859" lastFinishedPulling="2025-12-04 19:26:36.539047279 +0000 UTC m=+6458.494408325" observedRunningTime="2025-12-04 19:26:37.743493107 +0000 UTC m=+6459.698854173" watchObservedRunningTime="2025-12-04 19:26:37.752147801 +0000 UTC m=+6459.707508847" Dec 04 19:26:39 crc kubenswrapper[4733]: I1204 19:26:39.747116 4733 generic.go:334] "Generic (PLEG): container finished" podID="9595436c-1e86-4790-8ca4-3707f0d22b0c" containerID="7d37f601d15b0a66fc6301a369fbab32c1f52dfce3a3eb81b04d53cb4228476e" exitCode=0 Dec 04 19:26:39 crc kubenswrapper[4733]: I1204 19:26:39.747233 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-7jggv" event={"ID":"9595436c-1e86-4790-8ca4-3707f0d22b0c","Type":"ContainerDied","Data":"7d37f601d15b0a66fc6301a369fbab32c1f52dfce3a3eb81b04d53cb4228476e"} Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.181781 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.268240 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-config-data\") pod \"9595436c-1e86-4790-8ca4-3707f0d22b0c\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.268514 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-scripts\") pod \"9595436c-1e86-4790-8ca4-3707f0d22b0c\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.268618 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzbjs\" (UniqueName: \"kubernetes.io/projected/9595436c-1e86-4790-8ca4-3707f0d22b0c-kube-api-access-qzbjs\") pod \"9595436c-1e86-4790-8ca4-3707f0d22b0c\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.268675 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-combined-ca-bundle\") pod \"9595436c-1e86-4790-8ca4-3707f0d22b0c\" (UID: \"9595436c-1e86-4790-8ca4-3707f0d22b0c\") " Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.273806 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-scripts" (OuterVolumeSpecName: "scripts") pod "9595436c-1e86-4790-8ca4-3707f0d22b0c" (UID: "9595436c-1e86-4790-8ca4-3707f0d22b0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.274777 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9595436c-1e86-4790-8ca4-3707f0d22b0c-kube-api-access-qzbjs" (OuterVolumeSpecName: "kube-api-access-qzbjs") pod "9595436c-1e86-4790-8ca4-3707f0d22b0c" (UID: "9595436c-1e86-4790-8ca4-3707f0d22b0c"). InnerVolumeSpecName "kube-api-access-qzbjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.309733 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9595436c-1e86-4790-8ca4-3707f0d22b0c" (UID: "9595436c-1e86-4790-8ca4-3707f0d22b0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.310457 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-config-data" (OuterVolumeSpecName: "config-data") pod "9595436c-1e86-4790-8ca4-3707f0d22b0c" (UID: "9595436c-1e86-4790-8ca4-3707f0d22b0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.371552 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.371608 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzbjs\" (UniqueName: \"kubernetes.io/projected/9595436c-1e86-4790-8ca4-3707f0d22b0c-kube-api-access-qzbjs\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.371624 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.371935 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9595436c-1e86-4790-8ca4-3707f0d22b0c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.773726 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-7jggv" event={"ID":"9595436c-1e86-4790-8ca4-3707f0d22b0c","Type":"ContainerDied","Data":"4004a2eddc9316c32395d80d69de10bbb8726e0de0b0faac96a3dd185e3cc721"} Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.773830 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4004a2eddc9316c32395d80d69de10bbb8726e0de0b0faac96a3dd185e3cc721" Dec 04 19:26:41 crc kubenswrapper[4733]: I1204 19:26:41.773839 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-7jggv" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.198191 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.361983 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.362061 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.362118 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.363396 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.363505 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" gracePeriod=600 Dec 04 19:26:45 crc kubenswrapper[4733]: E1204 19:26:45.485735 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.794068 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 04 19:26:45 crc kubenswrapper[4733]: E1204 19:26:45.794820 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9595436c-1e86-4790-8ca4-3707f0d22b0c" containerName="aodh-db-sync" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.794861 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9595436c-1e86-4790-8ca4-3707f0d22b0c" containerName="aodh-db-sync" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.795150 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9595436c-1e86-4790-8ca4-3707f0d22b0c" containerName="aodh-db-sync" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.798474 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.807576 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.810630 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.816254 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.816460 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-kc27c" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.825182 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" exitCode=0 Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.825219 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08"} Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.825248 4733 scope.go:117] "RemoveContainer" containerID="353758af15d16464418cca9546b4d8b6ebe8b41cef4796d49135ac10e1a77f6b" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.825887 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:26:45 crc kubenswrapper[4733]: E1204 19:26:45.826134 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.866922 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr67n\" (UniqueName: \"kubernetes.io/projected/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-kube-api-access-dr67n\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.866964 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-scripts\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.867237 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-config-data\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.867376 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.969358 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-config-data\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.969446 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.969789 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr67n\" (UniqueName: \"kubernetes.io/projected/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-kube-api-access-dr67n\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.969822 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-scripts\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.980563 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:45 crc kubenswrapper[4733]: I1204 19:26:45.984018 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-scripts\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:46 crc kubenswrapper[4733]: I1204 19:26:45.999774 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr67n\" (UniqueName: \"kubernetes.io/projected/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-kube-api-access-dr67n\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:46 crc kubenswrapper[4733]: I1204 19:26:46.010465 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc95841e-bba4-46fa-ae27-1b956e0d3a0c-config-data\") pod \"aodh-0\" (UID: \"fc95841e-bba4-46fa-ae27-1b956e0d3a0c\") " pod="openstack/aodh-0" Dec 04 19:26:46 crc kubenswrapper[4733]: I1204 19:26:46.121361 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 19:26:46 crc kubenswrapper[4733]: I1204 19:26:46.628654 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 19:26:46 crc kubenswrapper[4733]: I1204 19:26:46.841883 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fc95841e-bba4-46fa-ae27-1b956e0d3a0c","Type":"ContainerStarted","Data":"a07c12c7252bc498b8a94cc0054ffb06f997a9c2e3fd4805f3c124f5eda6e583"} Dec 04 19:26:47 crc kubenswrapper[4733]: I1204 19:26:47.852739 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fc95841e-bba4-46fa-ae27-1b956e0d3a0c","Type":"ContainerStarted","Data":"1a3389a622e651acc98c5aeea371d5b1e5e3f3826026e4037bd434e96ccb03ab"} Dec 04 19:26:47 crc kubenswrapper[4733]: I1204 19:26:47.911595 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:47 crc kubenswrapper[4733]: I1204 19:26:47.911887 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="ceilometer-central-agent" containerID="cri-o://961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f" gracePeriod=30 Dec 04 19:26:47 crc kubenswrapper[4733]: I1204 19:26:47.912027 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="proxy-httpd" containerID="cri-o://fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb" gracePeriod=30 Dec 04 19:26:47 crc kubenswrapper[4733]: I1204 19:26:47.912069 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="sg-core" containerID="cri-o://341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5" gracePeriod=30 Dec 04 19:26:47 crc kubenswrapper[4733]: I1204 19:26:47.912103 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="ceilometer-notification-agent" containerID="cri-o://487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f" gracePeriod=30 Dec 04 19:26:48 crc kubenswrapper[4733]: I1204 19:26:48.863275 4733 generic.go:334] "Generic (PLEG): container finished" podID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerID="fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb" exitCode=0 Dec 04 19:26:48 crc kubenswrapper[4733]: I1204 19:26:48.863557 4733 generic.go:334] "Generic (PLEG): container finished" podID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerID="341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5" exitCode=2 Dec 04 19:26:48 crc kubenswrapper[4733]: I1204 19:26:48.863567 4733 generic.go:334] "Generic (PLEG): container finished" podID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerID="961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f" exitCode=0 Dec 04 19:26:48 crc kubenswrapper[4733]: I1204 19:26:48.863334 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerDied","Data":"fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb"} Dec 04 19:26:48 crc kubenswrapper[4733]: I1204 19:26:48.863596 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerDied","Data":"341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5"} Dec 04 19:26:48 crc kubenswrapper[4733]: I1204 19:26:48.863606 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerDied","Data":"961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f"} Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.501349 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.658467 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6brld\" (UniqueName: \"kubernetes.io/projected/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-kube-api-access-6brld\") pod \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.658586 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-run-httpd\") pod \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.658612 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-scripts\") pod \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.658674 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-log-httpd\") pod \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.658770 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-sg-core-conf-yaml\") pod \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.658830 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-config-data\") pod \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.659070 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" (UID: "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.658972 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-combined-ca-bundle\") pod \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\" (UID: \"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf\") " Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.659969 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.660076 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" (UID: "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.666461 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-scripts" (OuterVolumeSpecName: "scripts") pod "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" (UID: "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.666546 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-kube-api-access-6brld" (OuterVolumeSpecName: "kube-api-access-6brld") pod "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" (UID: "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf"). InnerVolumeSpecName "kube-api-access-6brld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.716583 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" (UID: "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.761293 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.761672 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6brld\" (UniqueName: \"kubernetes.io/projected/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-kube-api-access-6brld\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.761697 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.761706 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.777211 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" (UID: "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.819375 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-config-data" (OuterVolumeSpecName: "config-data") pod "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" (UID: "9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.863144 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.863181 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.882425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fc95841e-bba4-46fa-ae27-1b956e0d3a0c","Type":"ContainerStarted","Data":"f6cd99b7d2819db0a632985d4f6dce75a1e600401124528b411cfaf9468f7144"} Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.890143 4733 generic.go:334] "Generic (PLEG): container finished" podID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerID="487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f" exitCode=0 Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.890184 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerDied","Data":"487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f"} Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.890212 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf","Type":"ContainerDied","Data":"5faca64471e0b151a655805eed16f7387fab63c35e28b3c5118c3b362f6c381b"} Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.890234 4733 scope.go:117] "RemoveContainer" containerID="fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.890364 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.927066 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.935868 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.957785 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:49 crc kubenswrapper[4733]: E1204 19:26:49.958242 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="ceilometer-central-agent" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.958265 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="ceilometer-central-agent" Dec 04 19:26:49 crc kubenswrapper[4733]: E1204 19:26:49.958279 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="ceilometer-notification-agent" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.958288 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="ceilometer-notification-agent" Dec 04 19:26:49 crc kubenswrapper[4733]: E1204 19:26:49.958308 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="proxy-httpd" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.958325 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="proxy-httpd" Dec 04 19:26:49 crc kubenswrapper[4733]: E1204 19:26:49.958340 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="sg-core" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.958349 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="sg-core" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.958614 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="proxy-httpd" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.958640 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="sg-core" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.958650 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="ceilometer-central-agent" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.958670 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" containerName="ceilometer-notification-agent" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.960586 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.963374 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.963702 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 19:26:49 crc kubenswrapper[4733]: I1204 19:26:49.973310 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.067513 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-log-httpd\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.067568 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.067616 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.067675 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-scripts\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.067697 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-config-data\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.067756 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgnvr\" (UniqueName: \"kubernetes.io/projected/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-kube-api-access-fgnvr\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.068002 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-run-httpd\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.172741 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-scripts\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.172788 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-config-data\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.172827 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgnvr\" (UniqueName: \"kubernetes.io/projected/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-kube-api-access-fgnvr\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.172906 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-run-httpd\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.172965 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-log-httpd\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.172998 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.173038 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.173783 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-run-httpd\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.174110 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-log-httpd\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.177567 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-scripts\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.177719 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-config-data\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.180560 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.197641 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgnvr\" (UniqueName: \"kubernetes.io/projected/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-kube-api-access-fgnvr\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.199336 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.277641 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.282409 4733 scope.go:117] "RemoveContainer" containerID="341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.345758 4733 scope.go:117] "RemoveContainer" containerID="487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.349304 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf" path="/var/lib/kubelet/pods/9a3cbb1c-22cd-4e94-8e52-b9f63a5d13cf/volumes" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.443042 4733 scope.go:117] "RemoveContainer" containerID="961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.506140 4733 scope.go:117] "RemoveContainer" containerID="fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb" Dec 04 19:26:50 crc kubenswrapper[4733]: E1204 19:26:50.507331 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb\": container with ID starting with fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb not found: ID does not exist" containerID="fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.507365 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb"} err="failed to get container status \"fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb\": rpc error: code = NotFound desc = could not find container \"fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb\": container with ID starting with fe25c260e642a63e7e759209e6e50e1bba2d3d3c82b73aa821466bfbf4b832eb not found: ID does not exist" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.507546 4733 scope.go:117] "RemoveContainer" containerID="341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5" Dec 04 19:26:50 crc kubenswrapper[4733]: E1204 19:26:50.508154 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5\": container with ID starting with 341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5 not found: ID does not exist" containerID="341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.508211 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5"} err="failed to get container status \"341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5\": rpc error: code = NotFound desc = could not find container \"341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5\": container with ID starting with 341f2e69d0e25d9c0442f17c14bd98604006e9a5467e7212018489012da9b4f5 not found: ID does not exist" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.508230 4733 scope.go:117] "RemoveContainer" containerID="487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f" Dec 04 19:26:50 crc kubenswrapper[4733]: E1204 19:26:50.508642 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f\": container with ID starting with 487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f not found: ID does not exist" containerID="487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.508667 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f"} err="failed to get container status \"487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f\": rpc error: code = NotFound desc = could not find container \"487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f\": container with ID starting with 487b1a3f11ccd3af71aaf6457e50c05d0e06b54eddb27369b9fa961795b0034f not found: ID does not exist" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.508683 4733 scope.go:117] "RemoveContainer" containerID="961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f" Dec 04 19:26:50 crc kubenswrapper[4733]: E1204 19:26:50.509028 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f\": container with ID starting with 961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f not found: ID does not exist" containerID="961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.509075 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f"} err="failed to get container status \"961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f\": rpc error: code = NotFound desc = could not find container \"961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f\": container with ID starting with 961dc81ec2544762ca044ba0481481fb229f23e01958e0aa9218e698b50b9d5f not found: ID does not exist" Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.830236 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:26:50 crc kubenswrapper[4733]: W1204 19:26:50.839023 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b8bdf5f_a601_4a13_ad99_662ffd99a8ec.slice/crio-33b7d8bd5c0f616feca0f9adcbcfa79986cda28c2698808472f309a468d0b86e WatchSource:0}: Error finding container 33b7d8bd5c0f616feca0f9adcbcfa79986cda28c2698808472f309a468d0b86e: Status 404 returned error can't find the container with id 33b7d8bd5c0f616feca0f9adcbcfa79986cda28c2698808472f309a468d0b86e Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.907449 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fc95841e-bba4-46fa-ae27-1b956e0d3a0c","Type":"ContainerStarted","Data":"d857494e2278e45fd8c8c42542c840153ce7b5c888716ba7ee147a408a79a2d9"} Dec 04 19:26:50 crc kubenswrapper[4733]: I1204 19:26:50.909262 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerStarted","Data":"33b7d8bd5c0f616feca0f9adcbcfa79986cda28c2698808472f309a468d0b86e"} Dec 04 19:26:51 crc kubenswrapper[4733]: I1204 19:26:51.921343 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerStarted","Data":"f1440223c310cb27e80a1ccc0a9956c4ec5629064a522ec137cbd3ca941b0c56"} Dec 04 19:26:52 crc kubenswrapper[4733]: I1204 19:26:52.935848 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fc95841e-bba4-46fa-ae27-1b956e0d3a0c","Type":"ContainerStarted","Data":"d68f54a6185337e0d9ab8063b5a4d820235e0bd24d2ba81486eccd1d3cec42bf"} Dec 04 19:26:52 crc kubenswrapper[4733]: I1204 19:26:52.964293 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.285327968 podStartE2EDuration="7.964276579s" podCreationTimestamp="2025-12-04 19:26:45 +0000 UTC" firstStartedPulling="2025-12-04 19:26:46.636029165 +0000 UTC m=+6468.591390211" lastFinishedPulling="2025-12-04 19:26:52.314977776 +0000 UTC m=+6474.270338822" observedRunningTime="2025-12-04 19:26:52.960063285 +0000 UTC m=+6474.915424331" watchObservedRunningTime="2025-12-04 19:26:52.964276579 +0000 UTC m=+6474.919637625" Dec 04 19:26:53 crc kubenswrapper[4733]: I1204 19:26:53.945687 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerStarted","Data":"5ef7c8211a0ce807f8cd720902dfecb1aab5101ac4183f4e9779032efbd1408e"} Dec 04 19:26:53 crc kubenswrapper[4733]: I1204 19:26:53.946219 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerStarted","Data":"4ae3ce8e96bd1aa1eb69e0d866b75e83c87919d71df2257030dafe92f15930d6"} Dec 04 19:26:55 crc kubenswrapper[4733]: I1204 19:26:55.971376 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerStarted","Data":"077e52be012b1e529f02f03c98bb07eb9b0cdb6c10b988a5b0567ee9f845f26e"} Dec 04 19:26:55 crc kubenswrapper[4733]: I1204 19:26:55.973537 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.614690 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.071074563 podStartE2EDuration="9.614668007s" podCreationTimestamp="2025-12-04 19:26:49 +0000 UTC" firstStartedPulling="2025-12-04 19:26:50.842978413 +0000 UTC m=+6472.798339459" lastFinishedPulling="2025-12-04 19:26:55.386571857 +0000 UTC m=+6477.341932903" observedRunningTime="2025-12-04 19:26:56.010430292 +0000 UTC m=+6477.965791338" watchObservedRunningTime="2025-12-04 19:26:58.614668007 +0000 UTC m=+6480.570029053" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.622953 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-jqdr9"] Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.625016 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-jqdr9" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.633037 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-jqdr9"] Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.676984 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/529f9e6e-31b1-4901-90c2-47637c26fcc3-operator-scripts\") pod \"manila-db-create-jqdr9\" (UID: \"529f9e6e-31b1-4901-90c2-47637c26fcc3\") " pod="openstack/manila-db-create-jqdr9" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.677063 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4lcl\" (UniqueName: \"kubernetes.io/projected/529f9e6e-31b1-4901-90c2-47637c26fcc3-kube-api-access-b4lcl\") pod \"manila-db-create-jqdr9\" (UID: \"529f9e6e-31b1-4901-90c2-47637c26fcc3\") " pod="openstack/manila-db-create-jqdr9" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.723962 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-a076-account-create-update-lsnb7"] Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.725817 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.732784 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-a076-account-create-update-lsnb7"] Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.761244 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.778794 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/529f9e6e-31b1-4901-90c2-47637c26fcc3-operator-scripts\") pod \"manila-db-create-jqdr9\" (UID: \"529f9e6e-31b1-4901-90c2-47637c26fcc3\") " pod="openstack/manila-db-create-jqdr9" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.779305 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4lcl\" (UniqueName: \"kubernetes.io/projected/529f9e6e-31b1-4901-90c2-47637c26fcc3-kube-api-access-b4lcl\") pod \"manila-db-create-jqdr9\" (UID: \"529f9e6e-31b1-4901-90c2-47637c26fcc3\") " pod="openstack/manila-db-create-jqdr9" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.779501 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/529f9e6e-31b1-4901-90c2-47637c26fcc3-operator-scripts\") pod \"manila-db-create-jqdr9\" (UID: \"529f9e6e-31b1-4901-90c2-47637c26fcc3\") " pod="openstack/manila-db-create-jqdr9" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.803392 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4lcl\" (UniqueName: \"kubernetes.io/projected/529f9e6e-31b1-4901-90c2-47637c26fcc3-kube-api-access-b4lcl\") pod \"manila-db-create-jqdr9\" (UID: \"529f9e6e-31b1-4901-90c2-47637c26fcc3\") " pod="openstack/manila-db-create-jqdr9" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.881145 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8691a4ff-ced3-4098-a64d-c6d169d93eba-operator-scripts\") pod \"manila-a076-account-create-update-lsnb7\" (UID: \"8691a4ff-ced3-4098-a64d-c6d169d93eba\") " pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.881599 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxs6w\" (UniqueName: \"kubernetes.io/projected/8691a4ff-ced3-4098-a64d-c6d169d93eba-kube-api-access-lxs6w\") pod \"manila-a076-account-create-update-lsnb7\" (UID: \"8691a4ff-ced3-4098-a64d-c6d169d93eba\") " pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.945356 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-jqdr9" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.983514 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxs6w\" (UniqueName: \"kubernetes.io/projected/8691a4ff-ced3-4098-a64d-c6d169d93eba-kube-api-access-lxs6w\") pod \"manila-a076-account-create-update-lsnb7\" (UID: \"8691a4ff-ced3-4098-a64d-c6d169d93eba\") " pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.983668 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8691a4ff-ced3-4098-a64d-c6d169d93eba-operator-scripts\") pod \"manila-a076-account-create-update-lsnb7\" (UID: \"8691a4ff-ced3-4098-a64d-c6d169d93eba\") " pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:26:58 crc kubenswrapper[4733]: I1204 19:26:58.985061 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8691a4ff-ced3-4098-a64d-c6d169d93eba-operator-scripts\") pod \"manila-a076-account-create-update-lsnb7\" (UID: \"8691a4ff-ced3-4098-a64d-c6d169d93eba\") " pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:26:59 crc kubenswrapper[4733]: I1204 19:26:59.006480 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxs6w\" (UniqueName: \"kubernetes.io/projected/8691a4ff-ced3-4098-a64d-c6d169d93eba-kube-api-access-lxs6w\") pod \"manila-a076-account-create-update-lsnb7\" (UID: \"8691a4ff-ced3-4098-a64d-c6d169d93eba\") " pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:26:59 crc kubenswrapper[4733]: I1204 19:26:59.075158 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:26:59 crc kubenswrapper[4733]: I1204 19:26:59.340078 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:26:59 crc kubenswrapper[4733]: E1204 19:26:59.340575 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:26:59 crc kubenswrapper[4733]: I1204 19:26:59.480788 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-jqdr9"] Dec 04 19:26:59 crc kubenswrapper[4733]: W1204 19:26:59.484580 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod529f9e6e_31b1_4901_90c2_47637c26fcc3.slice/crio-7dca67cbe481ec52040926f209a276eadac8a7314e2ce2c7d9403aa0b660001e WatchSource:0}: Error finding container 7dca67cbe481ec52040926f209a276eadac8a7314e2ce2c7d9403aa0b660001e: Status 404 returned error can't find the container with id 7dca67cbe481ec52040926f209a276eadac8a7314e2ce2c7d9403aa0b660001e Dec 04 19:26:59 crc kubenswrapper[4733]: I1204 19:26:59.670889 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-a076-account-create-update-lsnb7"] Dec 04 19:27:00 crc kubenswrapper[4733]: I1204 19:27:00.024386 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-a076-account-create-update-lsnb7" event={"ID":"8691a4ff-ced3-4098-a64d-c6d169d93eba","Type":"ContainerStarted","Data":"4261675e3a05ddfe8f7078a08c53837038356425668c371d7b41c87164ba9d2d"} Dec 04 19:27:00 crc kubenswrapper[4733]: I1204 19:27:00.024715 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-a076-account-create-update-lsnb7" event={"ID":"8691a4ff-ced3-4098-a64d-c6d169d93eba","Type":"ContainerStarted","Data":"81a59cd2202cbf9596e47df02a95c75ed8b15cf327f0f47b211c1a29116cfa2b"} Dec 04 19:27:00 crc kubenswrapper[4733]: I1204 19:27:00.030056 4733 generic.go:334] "Generic (PLEG): container finished" podID="529f9e6e-31b1-4901-90c2-47637c26fcc3" containerID="ff3c098b1d71580667746addf9641c19911a6828a70aaa5093b72ad9c1a9ff52" exitCode=0 Dec 04 19:27:00 crc kubenswrapper[4733]: I1204 19:27:00.030096 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-jqdr9" event={"ID":"529f9e6e-31b1-4901-90c2-47637c26fcc3","Type":"ContainerDied","Data":"ff3c098b1d71580667746addf9641c19911a6828a70aaa5093b72ad9c1a9ff52"} Dec 04 19:27:00 crc kubenswrapper[4733]: I1204 19:27:00.030118 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-jqdr9" event={"ID":"529f9e6e-31b1-4901-90c2-47637c26fcc3","Type":"ContainerStarted","Data":"7dca67cbe481ec52040926f209a276eadac8a7314e2ce2c7d9403aa0b660001e"} Dec 04 19:27:00 crc kubenswrapper[4733]: I1204 19:27:00.051626 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-a076-account-create-update-lsnb7" podStartSLOduration=2.05160552 podStartE2EDuration="2.05160552s" podCreationTimestamp="2025-12-04 19:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:27:00.044298353 +0000 UTC m=+6481.999659409" watchObservedRunningTime="2025-12-04 19:27:00.05160552 +0000 UTC m=+6482.006966566" Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.040383 4733 generic.go:334] "Generic (PLEG): container finished" podID="8691a4ff-ced3-4098-a64d-c6d169d93eba" containerID="4261675e3a05ddfe8f7078a08c53837038356425668c371d7b41c87164ba9d2d" exitCode=0 Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.040434 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-a076-account-create-update-lsnb7" event={"ID":"8691a4ff-ced3-4098-a64d-c6d169d93eba","Type":"ContainerDied","Data":"4261675e3a05ddfe8f7078a08c53837038356425668c371d7b41c87164ba9d2d"} Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.512173 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-jqdr9" Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.553095 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/529f9e6e-31b1-4901-90c2-47637c26fcc3-operator-scripts\") pod \"529f9e6e-31b1-4901-90c2-47637c26fcc3\" (UID: \"529f9e6e-31b1-4901-90c2-47637c26fcc3\") " Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.553264 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4lcl\" (UniqueName: \"kubernetes.io/projected/529f9e6e-31b1-4901-90c2-47637c26fcc3-kube-api-access-b4lcl\") pod \"529f9e6e-31b1-4901-90c2-47637c26fcc3\" (UID: \"529f9e6e-31b1-4901-90c2-47637c26fcc3\") " Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.562733 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/529f9e6e-31b1-4901-90c2-47637c26fcc3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "529f9e6e-31b1-4901-90c2-47637c26fcc3" (UID: "529f9e6e-31b1-4901-90c2-47637c26fcc3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.567928 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/529f9e6e-31b1-4901-90c2-47637c26fcc3-kube-api-access-b4lcl" (OuterVolumeSpecName: "kube-api-access-b4lcl") pod "529f9e6e-31b1-4901-90c2-47637c26fcc3" (UID: "529f9e6e-31b1-4901-90c2-47637c26fcc3"). InnerVolumeSpecName "kube-api-access-b4lcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.655708 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/529f9e6e-31b1-4901-90c2-47637c26fcc3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:01 crc kubenswrapper[4733]: I1204 19:27:01.655740 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4lcl\" (UniqueName: \"kubernetes.io/projected/529f9e6e-31b1-4901-90c2-47637c26fcc3-kube-api-access-b4lcl\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.055093 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-jqdr9" Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.055488 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-jqdr9" event={"ID":"529f9e6e-31b1-4901-90c2-47637c26fcc3","Type":"ContainerDied","Data":"7dca67cbe481ec52040926f209a276eadac8a7314e2ce2c7d9403aa0b660001e"} Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.055535 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dca67cbe481ec52040926f209a276eadac8a7314e2ce2c7d9403aa0b660001e" Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.571754 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.576920 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxs6w\" (UniqueName: \"kubernetes.io/projected/8691a4ff-ced3-4098-a64d-c6d169d93eba-kube-api-access-lxs6w\") pod \"8691a4ff-ced3-4098-a64d-c6d169d93eba\" (UID: \"8691a4ff-ced3-4098-a64d-c6d169d93eba\") " Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.576996 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8691a4ff-ced3-4098-a64d-c6d169d93eba-operator-scripts\") pod \"8691a4ff-ced3-4098-a64d-c6d169d93eba\" (UID: \"8691a4ff-ced3-4098-a64d-c6d169d93eba\") " Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.578337 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8691a4ff-ced3-4098-a64d-c6d169d93eba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8691a4ff-ced3-4098-a64d-c6d169d93eba" (UID: "8691a4ff-ced3-4098-a64d-c6d169d93eba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.583796 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8691a4ff-ced3-4098-a64d-c6d169d93eba-kube-api-access-lxs6w" (OuterVolumeSpecName: "kube-api-access-lxs6w") pod "8691a4ff-ced3-4098-a64d-c6d169d93eba" (UID: "8691a4ff-ced3-4098-a64d-c6d169d93eba"). InnerVolumeSpecName "kube-api-access-lxs6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.679844 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxs6w\" (UniqueName: \"kubernetes.io/projected/8691a4ff-ced3-4098-a64d-c6d169d93eba-kube-api-access-lxs6w\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:02 crc kubenswrapper[4733]: I1204 19:27:02.679876 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8691a4ff-ced3-4098-a64d-c6d169d93eba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:03 crc kubenswrapper[4733]: I1204 19:27:03.070618 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-a076-account-create-update-lsnb7" event={"ID":"8691a4ff-ced3-4098-a64d-c6d169d93eba","Type":"ContainerDied","Data":"81a59cd2202cbf9596e47df02a95c75ed8b15cf327f0f47b211c1a29116cfa2b"} Dec 04 19:27:03 crc kubenswrapper[4733]: I1204 19:27:03.070656 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81a59cd2202cbf9596e47df02a95c75ed8b15cf327f0f47b211c1a29116cfa2b" Dec 04 19:27:03 crc kubenswrapper[4733]: I1204 19:27:03.070736 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-a076-account-create-update-lsnb7" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.085501 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-7hhh8"] Dec 04 19:27:04 crc kubenswrapper[4733]: E1204 19:27:04.086315 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8691a4ff-ced3-4098-a64d-c6d169d93eba" containerName="mariadb-account-create-update" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.086332 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8691a4ff-ced3-4098-a64d-c6d169d93eba" containerName="mariadb-account-create-update" Dec 04 19:27:04 crc kubenswrapper[4733]: E1204 19:27:04.086352 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529f9e6e-31b1-4901-90c2-47637c26fcc3" containerName="mariadb-database-create" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.086360 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="529f9e6e-31b1-4901-90c2-47637c26fcc3" containerName="mariadb-database-create" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.086626 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="529f9e6e-31b1-4901-90c2-47637c26fcc3" containerName="mariadb-database-create" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.086656 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8691a4ff-ced3-4098-a64d-c6d169d93eba" containerName="mariadb-account-create-update" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.087584 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.089664 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-tq7cr" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.090832 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.099529 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-7hhh8"] Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.207020 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg6f9\" (UniqueName: \"kubernetes.io/projected/58018da3-626d-4016-a9c9-4ff07a24ccf9-kube-api-access-sg6f9\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.207421 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-config-data\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.207517 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-combined-ca-bundle\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.207580 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-job-config-data\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.309436 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-config-data\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.309509 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-combined-ca-bundle\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.309548 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-job-config-data\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.309647 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg6f9\" (UniqueName: \"kubernetes.io/projected/58018da3-626d-4016-a9c9-4ff07a24ccf9-kube-api-access-sg6f9\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.314489 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-job-config-data\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.314718 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-config-data\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.314784 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-combined-ca-bundle\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.332199 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg6f9\" (UniqueName: \"kubernetes.io/projected/58018da3-626d-4016-a9c9-4ff07a24ccf9-kube-api-access-sg6f9\") pod \"manila-db-sync-7hhh8\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:04 crc kubenswrapper[4733]: I1204 19:27:04.425041 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:05 crc kubenswrapper[4733]: I1204 19:27:05.040188 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-lww2m"] Dec 04 19:27:05 crc kubenswrapper[4733]: I1204 19:27:05.051173 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-7ec3-account-create-update-s4czg"] Dec 04 19:27:05 crc kubenswrapper[4733]: I1204 19:27:05.066944 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-lww2m"] Dec 04 19:27:05 crc kubenswrapper[4733]: I1204 19:27:05.082262 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-7ec3-account-create-update-s4czg"] Dec 04 19:27:05 crc kubenswrapper[4733]: I1204 19:27:05.204245 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-7hhh8"] Dec 04 19:27:06 crc kubenswrapper[4733]: I1204 19:27:06.124143 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-7hhh8" event={"ID":"58018da3-626d-4016-a9c9-4ff07a24ccf9","Type":"ContainerStarted","Data":"d40da2988d8e5186c167f3ac0c2f8f61071fb73a9501f0a569d2429d5dc8b1f5"} Dec 04 19:27:06 crc kubenswrapper[4733]: I1204 19:27:06.362263 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6" path="/var/lib/kubelet/pods/9c0e84bb-07b0-4e93-b1eb-7170a0f35ec6/volumes" Dec 04 19:27:06 crc kubenswrapper[4733]: I1204 19:27:06.362912 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8ab8394-2a72-4898-a02b-a5bee23bdc6a" path="/var/lib/kubelet/pods/a8ab8394-2a72-4898-a02b-a5bee23bdc6a/volumes" Dec 04 19:27:12 crc kubenswrapper[4733]: I1204 19:27:12.048517 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-szccb"] Dec 04 19:27:12 crc kubenswrapper[4733]: I1204 19:27:12.059219 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-szccb"] Dec 04 19:27:12 crc kubenswrapper[4733]: I1204 19:27:12.210776 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-7hhh8" event={"ID":"58018da3-626d-4016-a9c9-4ff07a24ccf9","Type":"ContainerStarted","Data":"0822242418ead5fc50671a3368e47e07d7edd055a303357a076b0f92cb517633"} Dec 04 19:27:12 crc kubenswrapper[4733]: I1204 19:27:12.231736 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-7hhh8" podStartSLOduration=2.404595615 podStartE2EDuration="8.231719979s" podCreationTimestamp="2025-12-04 19:27:04 +0000 UTC" firstStartedPulling="2025-12-04 19:27:05.212917464 +0000 UTC m=+6487.168278500" lastFinishedPulling="2025-12-04 19:27:11.040041788 +0000 UTC m=+6492.995402864" observedRunningTime="2025-12-04 19:27:12.228600095 +0000 UTC m=+6494.183961141" watchObservedRunningTime="2025-12-04 19:27:12.231719979 +0000 UTC m=+6494.187081035" Dec 04 19:27:12 crc kubenswrapper[4733]: I1204 19:27:12.347561 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21a9d868-6c00-4586-b785-9efb6282c204" path="/var/lib/kubelet/pods/21a9d868-6c00-4586-b785-9efb6282c204/volumes" Dec 04 19:27:13 crc kubenswrapper[4733]: I1204 19:27:13.336773 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:27:13 crc kubenswrapper[4733]: E1204 19:27:13.337831 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:27:14 crc kubenswrapper[4733]: I1204 19:27:14.233285 4733 generic.go:334] "Generic (PLEG): container finished" podID="58018da3-626d-4016-a9c9-4ff07a24ccf9" containerID="0822242418ead5fc50671a3368e47e07d7edd055a303357a076b0f92cb517633" exitCode=0 Dec 04 19:27:14 crc kubenswrapper[4733]: I1204 19:27:14.233340 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-7hhh8" event={"ID":"58018da3-626d-4016-a9c9-4ff07a24ccf9","Type":"ContainerDied","Data":"0822242418ead5fc50671a3368e47e07d7edd055a303357a076b0f92cb517633"} Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.765545 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.871979 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-job-config-data\") pod \"58018da3-626d-4016-a9c9-4ff07a24ccf9\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.872096 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-config-data\") pod \"58018da3-626d-4016-a9c9-4ff07a24ccf9\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.872175 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg6f9\" (UniqueName: \"kubernetes.io/projected/58018da3-626d-4016-a9c9-4ff07a24ccf9-kube-api-access-sg6f9\") pod \"58018da3-626d-4016-a9c9-4ff07a24ccf9\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.872313 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-combined-ca-bundle\") pod \"58018da3-626d-4016-a9c9-4ff07a24ccf9\" (UID: \"58018da3-626d-4016-a9c9-4ff07a24ccf9\") " Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.878320 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58018da3-626d-4016-a9c9-4ff07a24ccf9-kube-api-access-sg6f9" (OuterVolumeSpecName: "kube-api-access-sg6f9") pod "58018da3-626d-4016-a9c9-4ff07a24ccf9" (UID: "58018da3-626d-4016-a9c9-4ff07a24ccf9"). InnerVolumeSpecName "kube-api-access-sg6f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.878768 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "58018da3-626d-4016-a9c9-4ff07a24ccf9" (UID: "58018da3-626d-4016-a9c9-4ff07a24ccf9"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.882882 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-config-data" (OuterVolumeSpecName: "config-data") pod "58018da3-626d-4016-a9c9-4ff07a24ccf9" (UID: "58018da3-626d-4016-a9c9-4ff07a24ccf9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.902368 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58018da3-626d-4016-a9c9-4ff07a24ccf9" (UID: "58018da3-626d-4016-a9c9-4ff07a24ccf9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.976663 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.976713 4733 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.976829 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58018da3-626d-4016-a9c9-4ff07a24ccf9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:15 crc kubenswrapper[4733]: I1204 19:27:15.976849 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg6f9\" (UniqueName: \"kubernetes.io/projected/58018da3-626d-4016-a9c9-4ff07a24ccf9-kube-api-access-sg6f9\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.256363 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-7hhh8" event={"ID":"58018da3-626d-4016-a9c9-4ff07a24ccf9","Type":"ContainerDied","Data":"d40da2988d8e5186c167f3ac0c2f8f61071fb73a9501f0a569d2429d5dc8b1f5"} Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.256999 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d40da2988d8e5186c167f3ac0c2f8f61071fb73a9501f0a569d2429d5dc8b1f5" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.257094 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-7hhh8" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.598868 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 04 19:27:16 crc kubenswrapper[4733]: E1204 19:27:16.599953 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58018da3-626d-4016-a9c9-4ff07a24ccf9" containerName="manila-db-sync" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.600025 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="58018da3-626d-4016-a9c9-4ff07a24ccf9" containerName="manila-db-sync" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.600284 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="58018da3-626d-4016-a9c9-4ff07a24ccf9" containerName="manila-db-sync" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.601560 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.604849 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.605101 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-tq7cr" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.605230 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.606137 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.616132 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.621028 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.625479 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.695001 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.701910 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.702245 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-config-data\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.702335 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.702541 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.702669 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-scripts\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.703164 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-ceph\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.703251 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.703344 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9fs4\" (UniqueName: \"kubernetes.io/projected/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-kube-api-access-w9fs4\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.716415 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.769098 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58d9d86f79-wbp6n"] Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.771751 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.781847 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58d9d86f79-wbp6n"] Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.804781 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-scripts\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.804862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.804881 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-config-data\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.804904 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.804956 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.804996 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805040 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-scripts\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805058 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s2lc\" (UniqueName: \"kubernetes.io/projected/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-kube-api-access-6s2lc\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805108 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-config-data\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805125 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805162 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-ceph\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805182 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805204 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805220 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9fs4\" (UniqueName: \"kubernetes.io/projected/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-kube-api-access-w9fs4\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.805651 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.806845 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.812255 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.817253 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-scripts\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.819359 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-ceph\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.821746 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-config-data\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.826740 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.833140 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9fs4\" (UniqueName: \"kubernetes.io/projected/d0ee1147-36ba-4d5e-95eb-17e8f8007e23-kube-api-access-w9fs4\") pod \"manila-share-share1-0\" (UID: \"d0ee1147-36ba-4d5e-95eb-17e8f8007e23\") " pod="openstack/manila-share-share1-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.907902 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-sb\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.908148 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.908292 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-scripts\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.908453 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pk74\" (UniqueName: \"kubernetes.io/projected/c087cdb3-7b59-4abf-be65-590d0ee4166b-kube-api-access-8pk74\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.908586 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-nb\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.908927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.909119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s2lc\" (UniqueName: \"kubernetes.io/projected/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-kube-api-access-6s2lc\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.909207 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-dns-svc\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.909341 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-config\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.909483 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-config-data\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.909588 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.910405 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.915322 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.917416 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.918315 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-config-data\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.918755 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-scripts\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.934518 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s2lc\" (UniqueName: \"kubernetes.io/projected/ccbe2eef-52a5-41b9-9000-bb0376d74bdc-kube-api-access-6s2lc\") pod \"manila-scheduler-0\" (UID: \"ccbe2eef-52a5-41b9-9000-bb0376d74bdc\") " pod="openstack/manila-scheduler-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.934933 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.937557 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.939680 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.945429 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 04 19:27:16 crc kubenswrapper[4733]: I1204 19:27:16.986241 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.014440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-nb\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.015368 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-nb\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.015936 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-dns-svc\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.016851 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-dns-svc\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.017079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-config\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.017743 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-config\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.018060 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-sb\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.018749 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-sb\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.020932 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pk74\" (UniqueName: \"kubernetes.io/projected/c087cdb3-7b59-4abf-be65-590d0ee4166b-kube-api-access-8pk74\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.036949 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pk74\" (UniqueName: \"kubernetes.io/projected/c087cdb3-7b59-4abf-be65-590d0ee4166b-kube-api-access-8pk74\") pod \"dnsmasq-dns-58d9d86f79-wbp6n\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.094188 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.124313 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-config-data\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.124454 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxp2n\" (UniqueName: \"kubernetes.io/projected/8b43f863-6942-4b7b-a4e5-9f6265204ad4-kube-api-access-gxp2n\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.124490 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.124575 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b43f863-6942-4b7b-a4e5-9f6265204ad4-etc-machine-id\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.125876 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-config-data-custom\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.125901 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-scripts\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.125927 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b43f863-6942-4b7b-a4e5-9f6265204ad4-logs\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.225311 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.228340 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-config-data\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.228408 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxp2n\" (UniqueName: \"kubernetes.io/projected/8b43f863-6942-4b7b-a4e5-9f6265204ad4-kube-api-access-gxp2n\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.228440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.228483 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b43f863-6942-4b7b-a4e5-9f6265204ad4-etc-machine-id\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.228684 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-config-data-custom\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.228736 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-scripts\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.228765 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b43f863-6942-4b7b-a4e5-9f6265204ad4-logs\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.229265 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b43f863-6942-4b7b-a4e5-9f6265204ad4-etc-machine-id\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.229506 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b43f863-6942-4b7b-a4e5-9f6265204ad4-logs\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.236131 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-scripts\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.236144 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.236152 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-config-data-custom\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.239529 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b43f863-6942-4b7b-a4e5-9f6265204ad4-config-data\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.260761 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxp2n\" (UniqueName: \"kubernetes.io/projected/8b43f863-6942-4b7b-a4e5-9f6265204ad4-kube-api-access-gxp2n\") pod \"manila-api-0\" (UID: \"8b43f863-6942-4b7b-a4e5-9f6265204ad4\") " pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.378261 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.656072 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.732407 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58d9d86f79-wbp6n"] Dec 04 19:27:17 crc kubenswrapper[4733]: W1204 19:27:17.747201 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc087cdb3_7b59_4abf_be65_590d0ee4166b.slice/crio-0aefae37840687ee1bd4c1a4f11f78c18269ca19a7b2b6b10677b37a024acc1d WatchSource:0}: Error finding container 0aefae37840687ee1bd4c1a4f11f78c18269ca19a7b2b6b10677b37a024acc1d: Status 404 returned error can't find the container with id 0aefae37840687ee1bd4c1a4f11f78c18269ca19a7b2b6b10677b37a024acc1d Dec 04 19:27:17 crc kubenswrapper[4733]: I1204 19:27:17.941284 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 04 19:27:17 crc kubenswrapper[4733]: W1204 19:27:17.952965 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccbe2eef_52a5_41b9_9000_bb0376d74bdc.slice/crio-a25195b514be6978783c61e50aa2676294f22d1e73ff8c19ce20a5f6463f46e8 WatchSource:0}: Error finding container a25195b514be6978783c61e50aa2676294f22d1e73ff8c19ce20a5f6463f46e8: Status 404 returned error can't find the container with id a25195b514be6978783c61e50aa2676294f22d1e73ff8c19ce20a5f6463f46e8 Dec 04 19:27:18 crc kubenswrapper[4733]: I1204 19:27:18.184289 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 04 19:27:18 crc kubenswrapper[4733]: W1204 19:27:18.223020 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b43f863_6942_4b7b_a4e5_9f6265204ad4.slice/crio-3302cdbdaf08bca368c33dfa7d69b4fabbfcf317a7dbd17f891e40a334d74241 WatchSource:0}: Error finding container 3302cdbdaf08bca368c33dfa7d69b4fabbfcf317a7dbd17f891e40a334d74241: Status 404 returned error can't find the container with id 3302cdbdaf08bca368c33dfa7d69b4fabbfcf317a7dbd17f891e40a334d74241 Dec 04 19:27:18 crc kubenswrapper[4733]: I1204 19:27:18.283339 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ccbe2eef-52a5-41b9-9000-bb0376d74bdc","Type":"ContainerStarted","Data":"a25195b514be6978783c61e50aa2676294f22d1e73ff8c19ce20a5f6463f46e8"} Dec 04 19:27:18 crc kubenswrapper[4733]: I1204 19:27:18.288575 4733 generic.go:334] "Generic (PLEG): container finished" podID="c087cdb3-7b59-4abf-be65-590d0ee4166b" containerID="88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81" exitCode=0 Dec 04 19:27:18 crc kubenswrapper[4733]: I1204 19:27:18.288642 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" event={"ID":"c087cdb3-7b59-4abf-be65-590d0ee4166b","Type":"ContainerDied","Data":"88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81"} Dec 04 19:27:18 crc kubenswrapper[4733]: I1204 19:27:18.288669 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" event={"ID":"c087cdb3-7b59-4abf-be65-590d0ee4166b","Type":"ContainerStarted","Data":"0aefae37840687ee1bd4c1a4f11f78c18269ca19a7b2b6b10677b37a024acc1d"} Dec 04 19:27:18 crc kubenswrapper[4733]: I1204 19:27:18.294180 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8b43f863-6942-4b7b-a4e5-9f6265204ad4","Type":"ContainerStarted","Data":"3302cdbdaf08bca368c33dfa7d69b4fabbfcf317a7dbd17f891e40a334d74241"} Dec 04 19:27:18 crc kubenswrapper[4733]: I1204 19:27:18.299169 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d0ee1147-36ba-4d5e-95eb-17e8f8007e23","Type":"ContainerStarted","Data":"ea5fe209d1c5ad3158c615be5eb853734c852aef7845adf55b766ddfdda773dd"} Dec 04 19:27:19 crc kubenswrapper[4733]: I1204 19:27:19.312284 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" event={"ID":"c087cdb3-7b59-4abf-be65-590d0ee4166b","Type":"ContainerStarted","Data":"2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762"} Dec 04 19:27:19 crc kubenswrapper[4733]: I1204 19:27:19.312876 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:19 crc kubenswrapper[4733]: I1204 19:27:19.316091 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8b43f863-6942-4b7b-a4e5-9f6265204ad4","Type":"ContainerStarted","Data":"766b99341f61a711da807902057cde82026916dc35e0dc2c1e84031e2abefa7b"} Dec 04 19:27:19 crc kubenswrapper[4733]: I1204 19:27:19.319129 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ccbe2eef-52a5-41b9-9000-bb0376d74bdc","Type":"ContainerStarted","Data":"8bf340d26403bc373771e507d9b99881f829b5f33e4ae3f6b30f3a44f4b2359d"} Dec 04 19:27:19 crc kubenswrapper[4733]: I1204 19:27:19.335396 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" podStartSLOduration=3.335382113 podStartE2EDuration="3.335382113s" podCreationTimestamp="2025-12-04 19:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:27:19.332992138 +0000 UTC m=+6501.288353184" watchObservedRunningTime="2025-12-04 19:27:19.335382113 +0000 UTC m=+6501.290743159" Dec 04 19:27:20 crc kubenswrapper[4733]: I1204 19:27:20.300206 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 19:27:20 crc kubenswrapper[4733]: I1204 19:27:20.361382 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8b43f863-6942-4b7b-a4e5-9f6265204ad4","Type":"ContainerStarted","Data":"bccf2ed7d65a791843561ca1c978db206869146682d07241040dace3d86b8f1f"} Dec 04 19:27:20 crc kubenswrapper[4733]: I1204 19:27:20.362582 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 04 19:27:20 crc kubenswrapper[4733]: I1204 19:27:20.362703 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ccbe2eef-52a5-41b9-9000-bb0376d74bdc","Type":"ContainerStarted","Data":"efa5321db2182ab2c8b3ef9aac7a6923b19bfe87600562dc7101bd14ebf11c48"} Dec 04 19:27:20 crc kubenswrapper[4733]: I1204 19:27:20.402648 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.402633125 podStartE2EDuration="4.402633125s" podCreationTimestamp="2025-12-04 19:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:27:20.375395047 +0000 UTC m=+6502.330756093" watchObservedRunningTime="2025-12-04 19:27:20.402633125 +0000 UTC m=+6502.357994171" Dec 04 19:27:21 crc kubenswrapper[4733]: I1204 19:27:21.530429 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.890533899 podStartE2EDuration="5.530404157s" podCreationTimestamp="2025-12-04 19:27:16 +0000 UTC" firstStartedPulling="2025-12-04 19:27:17.956983005 +0000 UTC m=+6499.912344061" lastFinishedPulling="2025-12-04 19:27:18.596853283 +0000 UTC m=+6500.552214319" observedRunningTime="2025-12-04 19:27:20.405733519 +0000 UTC m=+6502.361094595" watchObservedRunningTime="2025-12-04 19:27:21.530404157 +0000 UTC m=+6503.485765203" Dec 04 19:27:21 crc kubenswrapper[4733]: I1204 19:27:21.533113 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:21 crc kubenswrapper[4733]: I1204 19:27:21.533411 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="ceilometer-central-agent" containerID="cri-o://f1440223c310cb27e80a1ccc0a9956c4ec5629064a522ec137cbd3ca941b0c56" gracePeriod=30 Dec 04 19:27:21 crc kubenswrapper[4733]: I1204 19:27:21.534063 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="proxy-httpd" containerID="cri-o://077e52be012b1e529f02f03c98bb07eb9b0cdb6c10b988a5b0567ee9f845f26e" gracePeriod=30 Dec 04 19:27:21 crc kubenswrapper[4733]: I1204 19:27:21.534146 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="sg-core" containerID="cri-o://5ef7c8211a0ce807f8cd720902dfecb1aab5101ac4183f4e9779032efbd1408e" gracePeriod=30 Dec 04 19:27:21 crc kubenswrapper[4733]: I1204 19:27:21.534195 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="ceilometer-notification-agent" containerID="cri-o://4ae3ce8e96bd1aa1eb69e0d866b75e83c87919d71df2257030dafe92f15930d6" gracePeriod=30 Dec 04 19:27:22 crc kubenswrapper[4733]: I1204 19:27:22.384770 4733 generic.go:334] "Generic (PLEG): container finished" podID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerID="077e52be012b1e529f02f03c98bb07eb9b0cdb6c10b988a5b0567ee9f845f26e" exitCode=0 Dec 04 19:27:22 crc kubenswrapper[4733]: I1204 19:27:22.385121 4733 generic.go:334] "Generic (PLEG): container finished" podID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerID="5ef7c8211a0ce807f8cd720902dfecb1aab5101ac4183f4e9779032efbd1408e" exitCode=2 Dec 04 19:27:22 crc kubenswrapper[4733]: I1204 19:27:22.385133 4733 generic.go:334] "Generic (PLEG): container finished" podID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerID="f1440223c310cb27e80a1ccc0a9956c4ec5629064a522ec137cbd3ca941b0c56" exitCode=0 Dec 04 19:27:22 crc kubenswrapper[4733]: I1204 19:27:22.385383 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerDied","Data":"077e52be012b1e529f02f03c98bb07eb9b0cdb6c10b988a5b0567ee9f845f26e"} Dec 04 19:27:22 crc kubenswrapper[4733]: I1204 19:27:22.385414 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerDied","Data":"5ef7c8211a0ce807f8cd720902dfecb1aab5101ac4183f4e9779032efbd1408e"} Dec 04 19:27:22 crc kubenswrapper[4733]: I1204 19:27:22.385425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerDied","Data":"f1440223c310cb27e80a1ccc0a9956c4ec5629064a522ec137cbd3ca941b0c56"} Dec 04 19:27:25 crc kubenswrapper[4733]: I1204 19:27:25.430160 4733 generic.go:334] "Generic (PLEG): container finished" podID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerID="4ae3ce8e96bd1aa1eb69e0d866b75e83c87919d71df2257030dafe92f15930d6" exitCode=0 Dec 04 19:27:25 crc kubenswrapper[4733]: I1204 19:27:25.430255 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerDied","Data":"4ae3ce8e96bd1aa1eb69e0d866b75e83c87919d71df2257030dafe92f15930d6"} Dec 04 19:27:25 crc kubenswrapper[4733]: I1204 19:27:25.700624 4733 scope.go:117] "RemoveContainer" containerID="315c37c18e8368fb4803d151f7afbdb60d6429ff959bc96b58d3832fc05276d7" Dec 04 19:27:25 crc kubenswrapper[4733]: I1204 19:27:25.881752 4733 scope.go:117] "RemoveContainer" containerID="c7d4bf016c7eddc5eadae3a7b69d82654ea5c6e037dc4d50cc6ee226b99801bc" Dec 04 19:27:25 crc kubenswrapper[4733]: I1204 19:27:25.917150 4733 scope.go:117] "RemoveContainer" containerID="d6c619b4a7aa7643365dc04fd2a29d4f7daaa747f2c733d1329e4f509ddffe97" Dec 04 19:27:25 crc kubenswrapper[4733]: I1204 19:27:25.934577 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.043398 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-sg-core-conf-yaml\") pod \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.043467 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-combined-ca-bundle\") pod \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.043536 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgnvr\" (UniqueName: \"kubernetes.io/projected/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-kube-api-access-fgnvr\") pod \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.043557 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-scripts\") pod \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.043691 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-log-httpd\") pod \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.043753 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-config-data\") pod \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.043817 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-run-httpd\") pod \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\" (UID: \"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec\") " Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.044296 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" (UID: "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.044421 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" (UID: "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.051919 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-scripts" (OuterVolumeSpecName: "scripts") pod "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" (UID: "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.052255 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-kube-api-access-fgnvr" (OuterVolumeSpecName: "kube-api-access-fgnvr") pod "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" (UID: "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec"). InnerVolumeSpecName "kube-api-access-fgnvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.094450 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" (UID: "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.146258 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgnvr\" (UniqueName: \"kubernetes.io/projected/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-kube-api-access-fgnvr\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.146292 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.146305 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.146313 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.146324 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.155270 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" (UID: "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.181269 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-config-data" (OuterVolumeSpecName: "config-data") pod "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" (UID: "3b8bdf5f-a601-4a13-ad99-662ffd99a8ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.248535 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.248889 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.441024 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d0ee1147-36ba-4d5e-95eb-17e8f8007e23","Type":"ContainerStarted","Data":"ea238fc08ddd491f5f2adf5d339a092c80d55c2854154ba103e00e39e68811ee"} Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.446680 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b8bdf5f-a601-4a13-ad99-662ffd99a8ec","Type":"ContainerDied","Data":"33b7d8bd5c0f616feca0f9adcbcfa79986cda28c2698808472f309a468d0b86e"} Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.446728 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.446748 4733 scope.go:117] "RemoveContainer" containerID="077e52be012b1e529f02f03c98bb07eb9b0cdb6c10b988a5b0567ee9f845f26e" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.491161 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.503720 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.506448 4733 scope.go:117] "RemoveContainer" containerID="5ef7c8211a0ce807f8cd720902dfecb1aab5101ac4183f4e9779032efbd1408e" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.526975 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:26 crc kubenswrapper[4733]: E1204 19:27:26.527510 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="ceilometer-central-agent" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.527533 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="ceilometer-central-agent" Dec 04 19:27:26 crc kubenswrapper[4733]: E1204 19:27:26.527577 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="proxy-httpd" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.527586 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="proxy-httpd" Dec 04 19:27:26 crc kubenswrapper[4733]: E1204 19:27:26.527619 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="sg-core" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.527628 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="sg-core" Dec 04 19:27:26 crc kubenswrapper[4733]: E1204 19:27:26.527639 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="ceilometer-notification-agent" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.527647 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="ceilometer-notification-agent" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.527943 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="ceilometer-central-agent" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.527974 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="proxy-httpd" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.528083 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="sg-core" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.528125 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" containerName="ceilometer-notification-agent" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.543313 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.544094 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.547587 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.547985 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.565461 4733 scope.go:117] "RemoveContainer" containerID="4ae3ce8e96bd1aa1eb69e0d866b75e83c87919d71df2257030dafe92f15930d6" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.655952 4733 scope.go:117] "RemoveContainer" containerID="f1440223c310cb27e80a1ccc0a9956c4ec5629064a522ec137cbd3ca941b0c56" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.666609 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-log-httpd\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.666916 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq5f4\" (UniqueName: \"kubernetes.io/projected/7e16a807-8d10-4cea-8de7-e9fd73166c82-kube-api-access-mq5f4\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.666951 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-config-data\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.666975 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.667368 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.667561 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-run-httpd\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.667859 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-scripts\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.769383 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-log-httpd\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.769443 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq5f4\" (UniqueName: \"kubernetes.io/projected/7e16a807-8d10-4cea-8de7-e9fd73166c82-kube-api-access-mq5f4\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.769484 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-config-data\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.769515 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.769645 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.769694 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-run-httpd\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.769758 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-scripts\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.770778 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-run-httpd\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.771306 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-log-httpd\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.775773 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-scripts\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.775917 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.784237 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-config-data\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.787705 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.788258 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq5f4\" (UniqueName: \"kubernetes.io/projected/7e16a807-8d10-4cea-8de7-e9fd73166c82-kube-api-access-mq5f4\") pod \"ceilometer-0\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " pod="openstack/ceilometer-0" Dec 04 19:27:26 crc kubenswrapper[4733]: I1204 19:27:26.869048 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.098004 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.161645 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6"] Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.161981 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" podUID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" containerName="dnsmasq-dns" containerID="cri-o://226de081d3f6abc10ef577d59c4e903368f4e18a3940aed03dceaea641ecb565" gracePeriod=10 Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.226409 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.335693 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:27:27 crc kubenswrapper[4733]: E1204 19:27:27.336118 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.378487 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:27 crc kubenswrapper[4733]: W1204 19:27:27.383182 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e16a807_8d10_4cea_8de7_e9fd73166c82.slice/crio-a5ef782e93aa0abe5f313acb420de3ddd3c20bf54c9736fdeccc7f31673d3034 WatchSource:0}: Error finding container a5ef782e93aa0abe5f313acb420de3ddd3c20bf54c9736fdeccc7f31673d3034: Status 404 returned error can't find the container with id a5ef782e93aa0abe5f313acb420de3ddd3c20bf54c9736fdeccc7f31673d3034 Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.461784 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" containerID="226de081d3f6abc10ef577d59c4e903368f4e18a3940aed03dceaea641ecb565" exitCode=0 Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.462132 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" event={"ID":"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445","Type":"ContainerDied","Data":"226de081d3f6abc10ef577d59c4e903368f4e18a3940aed03dceaea641ecb565"} Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.465088 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d0ee1147-36ba-4d5e-95eb-17e8f8007e23","Type":"ContainerStarted","Data":"b3a5ac0a5ada0ddeb6459d6530e250573bc14243c7914a08913f9c795697c32c"} Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.474102 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerStarted","Data":"a5ef782e93aa0abe5f313acb420de3ddd3c20bf54c9736fdeccc7f31673d3034"} Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.504319 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.502675644 podStartE2EDuration="11.504300825s" podCreationTimestamp="2025-12-04 19:27:16 +0000 UTC" firstStartedPulling="2025-12-04 19:27:17.661984596 +0000 UTC m=+6499.617345642" lastFinishedPulling="2025-12-04 19:27:25.663609777 +0000 UTC m=+6507.618970823" observedRunningTime="2025-12-04 19:27:27.494067148 +0000 UTC m=+6509.449428204" watchObservedRunningTime="2025-12-04 19:27:27.504300825 +0000 UTC m=+6509.459661861" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.734919 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.793147 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnp9s\" (UniqueName: \"kubernetes.io/projected/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-kube-api-access-dnp9s\") pod \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.793383 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-config\") pod \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.793410 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-dns-svc\") pod \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.793506 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-nb\") pod \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.793589 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-sb\") pod \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\" (UID: \"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445\") " Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.805216 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-kube-api-access-dnp9s" (OuterVolumeSpecName: "kube-api-access-dnp9s") pod "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" (UID: "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445"). InnerVolumeSpecName "kube-api-access-dnp9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.867608 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" (UID: "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.868550 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" (UID: "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.871645 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" (UID: "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.891398 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-config" (OuterVolumeSpecName: "config") pod "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" (UID: "2d9d18c5-aa6c-411d-9f4b-9b34d55d4445"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.897369 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.897407 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.897417 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.897429 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:27 crc kubenswrapper[4733]: I1204 19:27:27.897439 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnp9s\" (UniqueName: \"kubernetes.io/projected/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445-kube-api-access-dnp9s\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:28 crc kubenswrapper[4733]: I1204 19:27:28.349828 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8bdf5f-a601-4a13-ad99-662ffd99a8ec" path="/var/lib/kubelet/pods/3b8bdf5f-a601-4a13-ad99-662ffd99a8ec/volumes" Dec 04 19:27:28 crc kubenswrapper[4733]: I1204 19:27:28.486063 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" event={"ID":"2d9d18c5-aa6c-411d-9f4b-9b34d55d4445","Type":"ContainerDied","Data":"81d9c3b238954b6de13b3558f3341fcfe1838b9ccfdc8ff6865d4552e56d108d"} Dec 04 19:27:28 crc kubenswrapper[4733]: I1204 19:27:28.485984 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6" Dec 04 19:27:28 crc kubenswrapper[4733]: I1204 19:27:28.486503 4733 scope.go:117] "RemoveContainer" containerID="226de081d3f6abc10ef577d59c4e903368f4e18a3940aed03dceaea641ecb565" Dec 04 19:27:28 crc kubenswrapper[4733]: I1204 19:27:28.492855 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerStarted","Data":"d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3"} Dec 04 19:27:28 crc kubenswrapper[4733]: I1204 19:27:28.529068 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6"] Dec 04 19:27:28 crc kubenswrapper[4733]: I1204 19:27:28.529380 4733 scope.go:117] "RemoveContainer" containerID="a560b36ba69efa48249f7ec05a8d0ec39bac0e40408c719a17a56e177d29dc38" Dec 04 19:27:28 crc kubenswrapper[4733]: I1204 19:27:28.539690 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d6cbfd8f5-9vdb6"] Dec 04 19:27:29 crc kubenswrapper[4733]: I1204 19:27:29.509158 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerStarted","Data":"8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807"} Dec 04 19:27:30 crc kubenswrapper[4733]: I1204 19:27:30.349852 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" path="/var/lib/kubelet/pods/2d9d18c5-aa6c-411d-9f4b-9b34d55d4445/volumes" Dec 04 19:27:30 crc kubenswrapper[4733]: I1204 19:27:30.423970 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:30 crc kubenswrapper[4733]: I1204 19:27:30.532374 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerStarted","Data":"b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd"} Dec 04 19:27:31 crc kubenswrapper[4733]: I1204 19:27:31.543283 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerStarted","Data":"8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def"} Dec 04 19:27:31 crc kubenswrapper[4733]: I1204 19:27:31.543464 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="ceilometer-central-agent" containerID="cri-o://d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3" gracePeriod=30 Dec 04 19:27:31 crc kubenswrapper[4733]: I1204 19:27:31.543703 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 19:27:31 crc kubenswrapper[4733]: I1204 19:27:31.543698 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="proxy-httpd" containerID="cri-o://8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def" gracePeriod=30 Dec 04 19:27:31 crc kubenswrapper[4733]: I1204 19:27:31.543707 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="sg-core" containerID="cri-o://b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd" gracePeriod=30 Dec 04 19:27:31 crc kubenswrapper[4733]: I1204 19:27:31.543727 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="ceilometer-notification-agent" containerID="cri-o://8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807" gracePeriod=30 Dec 04 19:27:31 crc kubenswrapper[4733]: I1204 19:27:31.563666 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.008462609 podStartE2EDuration="5.563643676s" podCreationTimestamp="2025-12-04 19:27:26 +0000 UTC" firstStartedPulling="2025-12-04 19:27:27.387361738 +0000 UTC m=+6509.342722784" lastFinishedPulling="2025-12-04 19:27:30.942542805 +0000 UTC m=+6512.897903851" observedRunningTime="2025-12-04 19:27:31.561826027 +0000 UTC m=+6513.517187083" watchObservedRunningTime="2025-12-04 19:27:31.563643676 +0000 UTC m=+6513.519004722" Dec 04 19:27:32 crc kubenswrapper[4733]: I1204 19:27:32.554847 4733 generic.go:334] "Generic (PLEG): container finished" podID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerID="8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def" exitCode=0 Dec 04 19:27:32 crc kubenswrapper[4733]: I1204 19:27:32.555116 4733 generic.go:334] "Generic (PLEG): container finished" podID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerID="b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd" exitCode=2 Dec 04 19:27:32 crc kubenswrapper[4733]: I1204 19:27:32.555130 4733 generic.go:334] "Generic (PLEG): container finished" podID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerID="8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807" exitCode=0 Dec 04 19:27:32 crc kubenswrapper[4733]: I1204 19:27:32.554937 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerDied","Data":"8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def"} Dec 04 19:27:32 crc kubenswrapper[4733]: I1204 19:27:32.555168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerDied","Data":"b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd"} Dec 04 19:27:32 crc kubenswrapper[4733]: I1204 19:27:32.555186 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerDied","Data":"8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807"} Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.323069 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.423281 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-scripts\") pod \"7e16a807-8d10-4cea-8de7-e9fd73166c82\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.423463 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-config-data\") pod \"7e16a807-8d10-4cea-8de7-e9fd73166c82\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.423525 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-combined-ca-bundle\") pod \"7e16a807-8d10-4cea-8de7-e9fd73166c82\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.423701 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-run-httpd\") pod \"7e16a807-8d10-4cea-8de7-e9fd73166c82\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.423733 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq5f4\" (UniqueName: \"kubernetes.io/projected/7e16a807-8d10-4cea-8de7-e9fd73166c82-kube-api-access-mq5f4\") pod \"7e16a807-8d10-4cea-8de7-e9fd73166c82\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.423831 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-sg-core-conf-yaml\") pod \"7e16a807-8d10-4cea-8de7-e9fd73166c82\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.423952 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-log-httpd\") pod \"7e16a807-8d10-4cea-8de7-e9fd73166c82\" (UID: \"7e16a807-8d10-4cea-8de7-e9fd73166c82\") " Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.424340 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7e16a807-8d10-4cea-8de7-e9fd73166c82" (UID: "7e16a807-8d10-4cea-8de7-e9fd73166c82"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.426296 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7e16a807-8d10-4cea-8de7-e9fd73166c82" (UID: "7e16a807-8d10-4cea-8de7-e9fd73166c82"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.433064 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e16a807-8d10-4cea-8de7-e9fd73166c82-kube-api-access-mq5f4" (OuterVolumeSpecName: "kube-api-access-mq5f4") pod "7e16a807-8d10-4cea-8de7-e9fd73166c82" (UID: "7e16a807-8d10-4cea-8de7-e9fd73166c82"). InnerVolumeSpecName "kube-api-access-mq5f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.436621 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-scripts" (OuterVolumeSpecName: "scripts") pod "7e16a807-8d10-4cea-8de7-e9fd73166c82" (UID: "7e16a807-8d10-4cea-8de7-e9fd73166c82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.457471 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7e16a807-8d10-4cea-8de7-e9fd73166c82" (UID: "7e16a807-8d10-4cea-8de7-e9fd73166c82"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.508736 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e16a807-8d10-4cea-8de7-e9fd73166c82" (UID: "7e16a807-8d10-4cea-8de7-e9fd73166c82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.525204 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-config-data" (OuterVolumeSpecName: "config-data") pod "7e16a807-8d10-4cea-8de7-e9fd73166c82" (UID: "7e16a807-8d10-4cea-8de7-e9fd73166c82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.526701 4733 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.526743 4733 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.526755 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.526765 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.526776 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e16a807-8d10-4cea-8de7-e9fd73166c82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.526787 4733 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e16a807-8d10-4cea-8de7-e9fd73166c82-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.526816 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq5f4\" (UniqueName: \"kubernetes.io/projected/7e16a807-8d10-4cea-8de7-e9fd73166c82-kube-api-access-mq5f4\") on node \"crc\" DevicePath \"\"" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.572553 4733 generic.go:334] "Generic (PLEG): container finished" podID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerID="d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3" exitCode=0 Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.572603 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerDied","Data":"d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3"} Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.572643 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e16a807-8d10-4cea-8de7-e9fd73166c82","Type":"ContainerDied","Data":"a5ef782e93aa0abe5f313acb420de3ddd3c20bf54c9736fdeccc7f31673d3034"} Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.572673 4733 scope.go:117] "RemoveContainer" containerID="8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.572905 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.599193 4733 scope.go:117] "RemoveContainer" containerID="b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.618586 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.631886 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.644827 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.645287 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="proxy-httpd" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645306 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="proxy-httpd" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.645324 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" containerName="dnsmasq-dns" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645331 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" containerName="dnsmasq-dns" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.645345 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" containerName="init" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645351 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" containerName="init" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.645369 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="ceilometer-central-agent" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645375 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="ceilometer-central-agent" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.645387 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="ceilometer-notification-agent" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645393 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="ceilometer-notification-agent" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.645418 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="sg-core" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645424 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="sg-core" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645606 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="ceilometer-notification-agent" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645629 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9d18c5-aa6c-411d-9f4b-9b34d55d4445" containerName="dnsmasq-dns" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645643 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="proxy-httpd" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645652 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="ceilometer-central-agent" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.645659 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" containerName="sg-core" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.652892 4733 scope.go:117] "RemoveContainer" containerID="8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.654105 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.657969 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.660703 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.661374 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.691751 4733 scope.go:117] "RemoveContainer" containerID="d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.716964 4733 scope.go:117] "RemoveContainer" containerID="8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.717455 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def\": container with ID starting with 8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def not found: ID does not exist" containerID="8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.717490 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def"} err="failed to get container status \"8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def\": rpc error: code = NotFound desc = could not find container \"8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def\": container with ID starting with 8faacb3237fc479d4f893bcd2e2cb5d5536fc9c8fa75239f4725dfa78e9b7def not found: ID does not exist" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.717511 4733 scope.go:117] "RemoveContainer" containerID="b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.717962 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd\": container with ID starting with b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd not found: ID does not exist" containerID="b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.718016 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd"} err="failed to get container status \"b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd\": rpc error: code = NotFound desc = could not find container \"b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd\": container with ID starting with b593e4faa74046d2a163ab66b86030cae4e5dffb658232432fc0671c4b02bdbd not found: ID does not exist" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.718081 4733 scope.go:117] "RemoveContainer" containerID="8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.718488 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807\": container with ID starting with 8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807 not found: ID does not exist" containerID="8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.718516 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807"} err="failed to get container status \"8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807\": rpc error: code = NotFound desc = could not find container \"8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807\": container with ID starting with 8d6ce198e98295c2206ed6b8a3b2b8d14e40d8156c584620d48f910b9eabb807 not found: ID does not exist" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.718531 4733 scope.go:117] "RemoveContainer" containerID="d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3" Dec 04 19:27:33 crc kubenswrapper[4733]: E1204 19:27:33.718866 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3\": container with ID starting with d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3 not found: ID does not exist" containerID="d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.718897 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3"} err="failed to get container status \"d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3\": rpc error: code = NotFound desc = could not find container \"d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3\": container with ID starting with d6edf32bc5e6e9c440f340b0ba7ffa70d23984e1186936357468a7736762fdc3 not found: ID does not exist" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.735484 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d793fc-05ac-4707-9457-e11c52ab66a9-run-httpd\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.735620 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.735655 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-config-data\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.735720 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.735747 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d793fc-05ac-4707-9457-e11c52ab66a9-log-httpd\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.735827 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxfgr\" (UniqueName: \"kubernetes.io/projected/46d793fc-05ac-4707-9457-e11c52ab66a9-kube-api-access-fxfgr\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.735897 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-scripts\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.838364 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d793fc-05ac-4707-9457-e11c52ab66a9-run-httpd\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.838471 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.838511 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-config-data\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.838598 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.838626 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d793fc-05ac-4707-9457-e11c52ab66a9-log-httpd\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.838700 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxfgr\" (UniqueName: \"kubernetes.io/projected/46d793fc-05ac-4707-9457-e11c52ab66a9-kube-api-access-fxfgr\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.838783 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-scripts\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.840697 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d793fc-05ac-4707-9457-e11c52ab66a9-log-httpd\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.840841 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d793fc-05ac-4707-9457-e11c52ab66a9-run-httpd\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.843627 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.843642 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-scripts\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.843643 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.847101 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d793fc-05ac-4707-9457-e11c52ab66a9-config-data\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.858831 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxfgr\" (UniqueName: \"kubernetes.io/projected/46d793fc-05ac-4707-9457-e11c52ab66a9-kube-api-access-fxfgr\") pod \"ceilometer-0\" (UID: \"46d793fc-05ac-4707-9457-e11c52ab66a9\") " pod="openstack/ceilometer-0" Dec 04 19:27:33 crc kubenswrapper[4733]: I1204 19:27:33.995563 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 19:27:34 crc kubenswrapper[4733]: I1204 19:27:34.355319 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e16a807-8d10-4cea-8de7-e9fd73166c82" path="/var/lib/kubelet/pods/7e16a807-8d10-4cea-8de7-e9fd73166c82/volumes" Dec 04 19:27:34 crc kubenswrapper[4733]: I1204 19:27:34.460956 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 19:27:34 crc kubenswrapper[4733]: W1204 19:27:34.463423 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46d793fc_05ac_4707_9457_e11c52ab66a9.slice/crio-ce2a88c0c1a3381ea95563cd2a6c16132e593fe6fff5b85eb8d8f936d38f1063 WatchSource:0}: Error finding container ce2a88c0c1a3381ea95563cd2a6c16132e593fe6fff5b85eb8d8f936d38f1063: Status 404 returned error can't find the container with id ce2a88c0c1a3381ea95563cd2a6c16132e593fe6fff5b85eb8d8f936d38f1063 Dec 04 19:27:34 crc kubenswrapper[4733]: I1204 19:27:34.582699 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d793fc-05ac-4707-9457-e11c52ab66a9","Type":"ContainerStarted","Data":"ce2a88c0c1a3381ea95563cd2a6c16132e593fe6fff5b85eb8d8f936d38f1063"} Dec 04 19:27:35 crc kubenswrapper[4733]: I1204 19:27:35.604722 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d793fc-05ac-4707-9457-e11c52ab66a9","Type":"ContainerStarted","Data":"6418a3a43a26a519bb20ad126c2a3e5fd8d829b38637166c4f8c49a1e5b1f009"} Dec 04 19:27:36 crc kubenswrapper[4733]: I1204 19:27:36.617942 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d793fc-05ac-4707-9457-e11c52ab66a9","Type":"ContainerStarted","Data":"3b4de41761caab6f58d2d467d26749668a16f406aaa41e9b6bd90d817e29950a"} Dec 04 19:27:36 crc kubenswrapper[4733]: I1204 19:27:36.618509 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d793fc-05ac-4707-9457-e11c52ab66a9","Type":"ContainerStarted","Data":"2cb7bc8eb27c52ed8412a603a35525a6cd01cdc32fbfb21a8a92529f063ee7ae"} Dec 04 19:27:36 crc kubenswrapper[4733]: I1204 19:27:36.987329 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 04 19:27:37 crc kubenswrapper[4733]: I1204 19:27:37.629849 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d793fc-05ac-4707-9457-e11c52ab66a9","Type":"ContainerStarted","Data":"77d4ab5c3ad75db797099e051bc1899cc43883d96bfb7cbe48f14aa9f4e5d96b"} Dec 04 19:27:37 crc kubenswrapper[4733]: I1204 19:27:37.630433 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 19:27:37 crc kubenswrapper[4733]: I1204 19:27:37.653548 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.798296543 podStartE2EDuration="4.653523845s" podCreationTimestamp="2025-12-04 19:27:33 +0000 UTC" firstStartedPulling="2025-12-04 19:27:34.466143428 +0000 UTC m=+6516.421504484" lastFinishedPulling="2025-12-04 19:27:37.32137074 +0000 UTC m=+6519.276731786" observedRunningTime="2025-12-04 19:27:37.645922269 +0000 UTC m=+6519.601283315" watchObservedRunningTime="2025-12-04 19:27:37.653523845 +0000 UTC m=+6519.608884891" Dec 04 19:27:38 crc kubenswrapper[4733]: I1204 19:27:38.576487 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 04 19:27:38 crc kubenswrapper[4733]: I1204 19:27:38.780631 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 04 19:27:38 crc kubenswrapper[4733]: I1204 19:27:38.932818 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 04 19:27:40 crc kubenswrapper[4733]: I1204 19:27:40.335878 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:27:40 crc kubenswrapper[4733]: E1204 19:27:40.336788 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.321886 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bqhkt"] Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.325826 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.335944 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bqhkt"] Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.380876 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af3a70e7-c70d-46bb-a133-1693b8da280a-utilities\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.380928 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af3a70e7-c70d-46bb-a133-1693b8da280a-catalog-content\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.381315 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfhjn\" (UniqueName: \"kubernetes.io/projected/af3a70e7-c70d-46bb-a133-1693b8da280a-kube-api-access-mfhjn\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.482773 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af3a70e7-c70d-46bb-a133-1693b8da280a-catalog-content\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.482896 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfhjn\" (UniqueName: \"kubernetes.io/projected/af3a70e7-c70d-46bb-a133-1693b8da280a-kube-api-access-mfhjn\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.483034 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af3a70e7-c70d-46bb-a133-1693b8da280a-utilities\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.483434 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af3a70e7-c70d-46bb-a133-1693b8da280a-catalog-content\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.483553 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af3a70e7-c70d-46bb-a133-1693b8da280a-utilities\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.515011 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfhjn\" (UniqueName: \"kubernetes.io/projected/af3a70e7-c70d-46bb-a133-1693b8da280a-kube-api-access-mfhjn\") pod \"community-operators-bqhkt\" (UID: \"af3a70e7-c70d-46bb-a133-1693b8da280a\") " pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:49 crc kubenswrapper[4733]: I1204 19:27:49.662033 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:27:50 crc kubenswrapper[4733]: I1204 19:27:50.252825 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bqhkt"] Dec 04 19:27:50 crc kubenswrapper[4733]: W1204 19:27:50.264350 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf3a70e7_c70d_46bb_a133_1693b8da280a.slice/crio-f4de532b8b41c93dfdf7cff3f86d8830cc4c8ca6e2aabde7b5b6da9cb1a9c2e2 WatchSource:0}: Error finding container f4de532b8b41c93dfdf7cff3f86d8830cc4c8ca6e2aabde7b5b6da9cb1a9c2e2: Status 404 returned error can't find the container with id f4de532b8b41c93dfdf7cff3f86d8830cc4c8ca6e2aabde7b5b6da9cb1a9c2e2 Dec 04 19:27:50 crc kubenswrapper[4733]: I1204 19:27:50.778738 4733 generic.go:334] "Generic (PLEG): container finished" podID="af3a70e7-c70d-46bb-a133-1693b8da280a" containerID="931b202d2052d0a95098cede93d15f4da41d9b727efa134b97405cc468c43add" exitCode=0 Dec 04 19:27:50 crc kubenswrapper[4733]: I1204 19:27:50.778844 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqhkt" event={"ID":"af3a70e7-c70d-46bb-a133-1693b8da280a","Type":"ContainerDied","Data":"931b202d2052d0a95098cede93d15f4da41d9b727efa134b97405cc468c43add"} Dec 04 19:27:50 crc kubenswrapper[4733]: I1204 19:27:50.779196 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqhkt" event={"ID":"af3a70e7-c70d-46bb-a133-1693b8da280a","Type":"ContainerStarted","Data":"f4de532b8b41c93dfdf7cff3f86d8830cc4c8ca6e2aabde7b5b6da9cb1a9c2e2"} Dec 04 19:27:55 crc kubenswrapper[4733]: I1204 19:27:55.335935 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:27:55 crc kubenswrapper[4733]: E1204 19:27:55.336782 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:27:56 crc kubenswrapper[4733]: I1204 19:27:56.842120 4733 generic.go:334] "Generic (PLEG): container finished" podID="af3a70e7-c70d-46bb-a133-1693b8da280a" containerID="7902c1182a737df724fa063529170a3a536666ab2045f824d86e36d138f6f821" exitCode=0 Dec 04 19:27:56 crc kubenswrapper[4733]: I1204 19:27:56.842180 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqhkt" event={"ID":"af3a70e7-c70d-46bb-a133-1693b8da280a","Type":"ContainerDied","Data":"7902c1182a737df724fa063529170a3a536666ab2045f824d86e36d138f6f821"} Dec 04 19:27:59 crc kubenswrapper[4733]: I1204 19:27:59.876950 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqhkt" event={"ID":"af3a70e7-c70d-46bb-a133-1693b8da280a","Type":"ContainerStarted","Data":"04add300c8c75cbc4cce9f7c0c5f753c338166c007124262af22dfef1bf0ecbe"} Dec 04 19:27:59 crc kubenswrapper[4733]: I1204 19:27:59.896249 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bqhkt" podStartSLOduration=2.203477371 podStartE2EDuration="10.896230949s" podCreationTimestamp="2025-12-04 19:27:49 +0000 UTC" firstStartedPulling="2025-12-04 19:27:50.781337829 +0000 UTC m=+6532.736698875" lastFinishedPulling="2025-12-04 19:27:59.474091407 +0000 UTC m=+6541.429452453" observedRunningTime="2025-12-04 19:27:59.895043477 +0000 UTC m=+6541.850404543" watchObservedRunningTime="2025-12-04 19:27:59.896230949 +0000 UTC m=+6541.851591995" Dec 04 19:28:04 crc kubenswrapper[4733]: I1204 19:28:04.001933 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 19:28:06 crc kubenswrapper[4733]: I1204 19:28:06.336197 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:28:06 crc kubenswrapper[4733]: E1204 19:28:06.336761 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:28:09 crc kubenswrapper[4733]: I1204 19:28:09.663246 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:28:09 crc kubenswrapper[4733]: I1204 19:28:09.663733 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:28:09 crc kubenswrapper[4733]: I1204 19:28:09.726160 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.060711 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bqhkt" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.141755 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bqhkt"] Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.212911 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cj6ms"] Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.213178 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cj6ms" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerName="registry-server" containerID="cri-o://0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2" gracePeriod=2 Dec 04 19:28:10 crc kubenswrapper[4733]: E1204 19:28:10.333568 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72e0892e_4912_407e_b6ee_438c98fb6d8a.slice/crio-0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72e0892e_4912_407e_b6ee_438c98fb6d8a.slice/crio-conmon-0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2.scope\": RecentStats: unable to find data in memory cache]" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.753541 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cj6ms" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.863867 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-utilities\") pod \"72e0892e-4912-407e-b6ee-438c98fb6d8a\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.864093 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-catalog-content\") pod \"72e0892e-4912-407e-b6ee-438c98fb6d8a\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.864118 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvdjm\" (UniqueName: \"kubernetes.io/projected/72e0892e-4912-407e-b6ee-438c98fb6d8a-kube-api-access-mvdjm\") pod \"72e0892e-4912-407e-b6ee-438c98fb6d8a\" (UID: \"72e0892e-4912-407e-b6ee-438c98fb6d8a\") " Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.864371 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-utilities" (OuterVolumeSpecName: "utilities") pod "72e0892e-4912-407e-b6ee-438c98fb6d8a" (UID: "72e0892e-4912-407e-b6ee-438c98fb6d8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.864548 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.871592 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e0892e-4912-407e-b6ee-438c98fb6d8a-kube-api-access-mvdjm" (OuterVolumeSpecName: "kube-api-access-mvdjm") pod "72e0892e-4912-407e-b6ee-438c98fb6d8a" (UID: "72e0892e-4912-407e-b6ee-438c98fb6d8a"). InnerVolumeSpecName "kube-api-access-mvdjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.918208 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72e0892e-4912-407e-b6ee-438c98fb6d8a" (UID: "72e0892e-4912-407e-b6ee-438c98fb6d8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.966769 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e0892e-4912-407e-b6ee-438c98fb6d8a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:10 crc kubenswrapper[4733]: I1204 19:28:10.966816 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvdjm\" (UniqueName: \"kubernetes.io/projected/72e0892e-4912-407e-b6ee-438c98fb6d8a-kube-api-access-mvdjm\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.004606 4733 generic.go:334] "Generic (PLEG): container finished" podID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerID="0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2" exitCode=0 Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.004665 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cj6ms" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.004717 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cj6ms" event={"ID":"72e0892e-4912-407e-b6ee-438c98fb6d8a","Type":"ContainerDied","Data":"0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2"} Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.004919 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cj6ms" event={"ID":"72e0892e-4912-407e-b6ee-438c98fb6d8a","Type":"ContainerDied","Data":"3a541950764629e12490f6bda2fb6e0c7103e86a63871431327713af913cd60f"} Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.004954 4733 scope.go:117] "RemoveContainer" containerID="0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.035743 4733 scope.go:117] "RemoveContainer" containerID="d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.058248 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cj6ms"] Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.069539 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cj6ms"] Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.089024 4733 scope.go:117] "RemoveContainer" containerID="e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.114282 4733 scope.go:117] "RemoveContainer" containerID="0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2" Dec 04 19:28:11 crc kubenswrapper[4733]: E1204 19:28:11.114946 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2\": container with ID starting with 0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2 not found: ID does not exist" containerID="0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.114978 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2"} err="failed to get container status \"0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2\": rpc error: code = NotFound desc = could not find container \"0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2\": container with ID starting with 0d316599f5636d9a7522fe8e914d68833ad736e9c52bb386ca38ffe502f726d2 not found: ID does not exist" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.114999 4733 scope.go:117] "RemoveContainer" containerID="d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee" Dec 04 19:28:11 crc kubenswrapper[4733]: E1204 19:28:11.115415 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee\": container with ID starting with d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee not found: ID does not exist" containerID="d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.115455 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee"} err="failed to get container status \"d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee\": rpc error: code = NotFound desc = could not find container \"d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee\": container with ID starting with d6bc6fbd7e4819762b53e3aeaaf5fcd8448dd0bacb951dc1b09f1ab47dab47ee not found: ID does not exist" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.115482 4733 scope.go:117] "RemoveContainer" containerID="e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692" Dec 04 19:28:11 crc kubenswrapper[4733]: E1204 19:28:11.115725 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692\": container with ID starting with e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692 not found: ID does not exist" containerID="e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692" Dec 04 19:28:11 crc kubenswrapper[4733]: I1204 19:28:11.115748 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692"} err="failed to get container status \"e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692\": rpc error: code = NotFound desc = could not find container \"e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692\": container with ID starting with e8a70687c52be3c61e7391bf86768356d40ac5a974ce03d3260598816eb06692 not found: ID does not exist" Dec 04 19:28:12 crc kubenswrapper[4733]: I1204 19:28:12.355720 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" path="/var/lib/kubelet/pods/72e0892e-4912-407e-b6ee-438c98fb6d8a/volumes" Dec 04 19:28:21 crc kubenswrapper[4733]: I1204 19:28:21.335477 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:28:21 crc kubenswrapper[4733]: E1204 19:28:21.336334 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.883313 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9cfcfbc7-xh59n"] Dec 04 19:28:23 crc kubenswrapper[4733]: E1204 19:28:23.884608 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerName="extract-content" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.884626 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerName="extract-content" Dec 04 19:28:23 crc kubenswrapper[4733]: E1204 19:28:23.884642 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerName="registry-server" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.884649 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerName="registry-server" Dec 04 19:28:23 crc kubenswrapper[4733]: E1204 19:28:23.884685 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerName="extract-utilities" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.884694 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerName="extract-utilities" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.884969 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e0892e-4912-407e-b6ee-438c98fb6d8a" containerName="registry-server" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.886584 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.888812 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.904585 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9cfcfbc7-xh59n"] Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.957923 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.958038 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.958080 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-openstack-cell1\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.958111 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-config\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.958184 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gjd6\" (UniqueName: \"kubernetes.io/projected/e7463837-6e34-4089-bd43-8912ec048984-kube-api-access-5gjd6\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:23 crc kubenswrapper[4733]: I1204 19:28:23.958212 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-dns-svc\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.060703 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-config\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.060818 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gjd6\" (UniqueName: \"kubernetes.io/projected/e7463837-6e34-4089-bd43-8912ec048984-kube-api-access-5gjd6\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.060866 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-dns-svc\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.060927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.061024 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.061069 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-openstack-cell1\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.061993 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-config\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.062013 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-openstack-cell1\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.062329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.062499 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.062726 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-dns-svc\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.090064 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gjd6\" (UniqueName: \"kubernetes.io/projected/e7463837-6e34-4089-bd43-8912ec048984-kube-api-access-5gjd6\") pod \"dnsmasq-dns-5c9cfcfbc7-xh59n\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.210768 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:24 crc kubenswrapper[4733]: I1204 19:28:24.673141 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9cfcfbc7-xh59n"] Dec 04 19:28:25 crc kubenswrapper[4733]: I1204 19:28:25.163081 4733 generic.go:334] "Generic (PLEG): container finished" podID="e7463837-6e34-4089-bd43-8912ec048984" containerID="b88862e5b704236f458bfb9308dd8cf17815aac1abb731dc679367ed601fe860" exitCode=0 Dec 04 19:28:25 crc kubenswrapper[4733]: I1204 19:28:25.164082 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" event={"ID":"e7463837-6e34-4089-bd43-8912ec048984","Type":"ContainerDied","Data":"b88862e5b704236f458bfb9308dd8cf17815aac1abb731dc679367ed601fe860"} Dec 04 19:28:25 crc kubenswrapper[4733]: I1204 19:28:25.164552 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" event={"ID":"e7463837-6e34-4089-bd43-8912ec048984","Type":"ContainerStarted","Data":"3ec79b28d05d511cc3a1fda8d8443fd579f9f6c8d0639a7e8287d6119aeb1112"} Dec 04 19:28:26 crc kubenswrapper[4733]: I1204 19:28:26.177945 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" event={"ID":"e7463837-6e34-4089-bd43-8912ec048984","Type":"ContainerStarted","Data":"c1e58f880faa02d2aa18227d05a3d2292ccc8e244b975d5a514d51904264da56"} Dec 04 19:28:26 crc kubenswrapper[4733]: I1204 19:28:26.178410 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:26 crc kubenswrapper[4733]: I1204 19:28:26.197593 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" podStartSLOduration=3.197567085 podStartE2EDuration="3.197567085s" podCreationTimestamp="2025-12-04 19:28:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:28:26.196471444 +0000 UTC m=+6568.151832510" watchObservedRunningTime="2025-12-04 19:28:26.197567085 +0000 UTC m=+6568.152928131" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.213084 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.285134 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58d9d86f79-wbp6n"] Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.285373 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" podUID="c087cdb3-7b59-4abf-be65-590d0ee4166b" containerName="dnsmasq-dns" containerID="cri-o://2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762" gracePeriod=10 Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.336232 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:28:34 crc kubenswrapper[4733]: E1204 19:28:34.336518 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.440128 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d959b8c7-kxj9z"] Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.442457 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.459846 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d959b8c7-kxj9z"] Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.498655 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-config\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.498756 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrk2v\" (UniqueName: \"kubernetes.io/projected/fbe1ced6-e57e-4add-8c40-a56fd21e055b-kube-api-access-vrk2v\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.498851 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-ovsdbserver-sb\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.498881 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-openstack-cell1\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.498968 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-ovsdbserver-nb\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.499052 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-dns-svc\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.604361 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-config\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.604457 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrk2v\" (UniqueName: \"kubernetes.io/projected/fbe1ced6-e57e-4add-8c40-a56fd21e055b-kube-api-access-vrk2v\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.604529 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-ovsdbserver-sb\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.604550 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-openstack-cell1\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.604625 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-ovsdbserver-nb\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.604690 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-dns-svc\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.605466 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-config\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.605531 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-dns-svc\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.606148 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-openstack-cell1\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.606309 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-ovsdbserver-sb\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.606519 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbe1ced6-e57e-4add-8c40-a56fd21e055b-ovsdbserver-nb\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.629821 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrk2v\" (UniqueName: \"kubernetes.io/projected/fbe1ced6-e57e-4add-8c40-a56fd21e055b-kube-api-access-vrk2v\") pod \"dnsmasq-dns-d959b8c7-kxj9z\" (UID: \"fbe1ced6-e57e-4add-8c40-a56fd21e055b\") " pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.781541 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:34 crc kubenswrapper[4733]: I1204 19:28:34.917121 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.114947 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pk74\" (UniqueName: \"kubernetes.io/projected/c087cdb3-7b59-4abf-be65-590d0ee4166b-kube-api-access-8pk74\") pod \"c087cdb3-7b59-4abf-be65-590d0ee4166b\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.115333 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-config\") pod \"c087cdb3-7b59-4abf-be65-590d0ee4166b\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.115455 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-sb\") pod \"c087cdb3-7b59-4abf-be65-590d0ee4166b\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.115577 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-nb\") pod \"c087cdb3-7b59-4abf-be65-590d0ee4166b\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.115620 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-dns-svc\") pod \"c087cdb3-7b59-4abf-be65-590d0ee4166b\" (UID: \"c087cdb3-7b59-4abf-be65-590d0ee4166b\") " Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.126605 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c087cdb3-7b59-4abf-be65-590d0ee4166b-kube-api-access-8pk74" (OuterVolumeSpecName: "kube-api-access-8pk74") pod "c087cdb3-7b59-4abf-be65-590d0ee4166b" (UID: "c087cdb3-7b59-4abf-be65-590d0ee4166b"). InnerVolumeSpecName "kube-api-access-8pk74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.196649 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c087cdb3-7b59-4abf-be65-590d0ee4166b" (UID: "c087cdb3-7b59-4abf-be65-590d0ee4166b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.198643 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-config" (OuterVolumeSpecName: "config") pod "c087cdb3-7b59-4abf-be65-590d0ee4166b" (UID: "c087cdb3-7b59-4abf-be65-590d0ee4166b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.217869 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pk74\" (UniqueName: \"kubernetes.io/projected/c087cdb3-7b59-4abf-be65-590d0ee4166b-kube-api-access-8pk74\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.217900 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.217911 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.219694 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c087cdb3-7b59-4abf-be65-590d0ee4166b" (UID: "c087cdb3-7b59-4abf-be65-590d0ee4166b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.222358 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c087cdb3-7b59-4abf-be65-590d0ee4166b" (UID: "c087cdb3-7b59-4abf-be65-590d0ee4166b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.287522 4733 generic.go:334] "Generic (PLEG): container finished" podID="c087cdb3-7b59-4abf-be65-590d0ee4166b" containerID="2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762" exitCode=0 Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.287579 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.287584 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" event={"ID":"c087cdb3-7b59-4abf-be65-590d0ee4166b","Type":"ContainerDied","Data":"2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762"} Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.290603 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d9d86f79-wbp6n" event={"ID":"c087cdb3-7b59-4abf-be65-590d0ee4166b","Type":"ContainerDied","Data":"0aefae37840687ee1bd4c1a4f11f78c18269ca19a7b2b6b10677b37a024acc1d"} Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.290629 4733 scope.go:117] "RemoveContainer" containerID="2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.322004 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.322041 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c087cdb3-7b59-4abf-be65-590d0ee4166b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.329097 4733 scope.go:117] "RemoveContainer" containerID="88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.334784 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58d9d86f79-wbp6n"] Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.355092 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58d9d86f79-wbp6n"] Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.362835 4733 scope.go:117] "RemoveContainer" containerID="2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762" Dec 04 19:28:35 crc kubenswrapper[4733]: E1204 19:28:35.363319 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762\": container with ID starting with 2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762 not found: ID does not exist" containerID="2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.363379 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762"} err="failed to get container status \"2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762\": rpc error: code = NotFound desc = could not find container \"2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762\": container with ID starting with 2fa14dc38879c00af5ce94f4a98be48f1f6039dabac03651ab45c3b08a65e762 not found: ID does not exist" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.363404 4733 scope.go:117] "RemoveContainer" containerID="88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81" Dec 04 19:28:35 crc kubenswrapper[4733]: E1204 19:28:35.363733 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81\": container with ID starting with 88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81 not found: ID does not exist" containerID="88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.363785 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81"} err="failed to get container status \"88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81\": rpc error: code = NotFound desc = could not find container \"88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81\": container with ID starting with 88f6cbbf57defb608273f53af77a0ef93c6efc00d992b4abae46d44c533d4f81 not found: ID does not exist" Dec 04 19:28:35 crc kubenswrapper[4733]: I1204 19:28:35.549937 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d959b8c7-kxj9z"] Dec 04 19:28:36 crc kubenswrapper[4733]: I1204 19:28:36.309145 4733 generic.go:334] "Generic (PLEG): container finished" podID="fbe1ced6-e57e-4add-8c40-a56fd21e055b" containerID="61c4d8d76c16f6938017203527bb4be8e76052cfdfa56470be027d8dbd57cbe6" exitCode=0 Dec 04 19:28:36 crc kubenswrapper[4733]: I1204 19:28:36.310623 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" event={"ID":"fbe1ced6-e57e-4add-8c40-a56fd21e055b","Type":"ContainerDied","Data":"61c4d8d76c16f6938017203527bb4be8e76052cfdfa56470be027d8dbd57cbe6"} Dec 04 19:28:36 crc kubenswrapper[4733]: I1204 19:28:36.310671 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" event={"ID":"fbe1ced6-e57e-4add-8c40-a56fd21e055b","Type":"ContainerStarted","Data":"0eec2cd96d78db377146d7b5fc50b036db7b4f4ed43498e8eec35bdef4e3838b"} Dec 04 19:28:36 crc kubenswrapper[4733]: I1204 19:28:36.363303 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c087cdb3-7b59-4abf-be65-590d0ee4166b" path="/var/lib/kubelet/pods/c087cdb3-7b59-4abf-be65-590d0ee4166b/volumes" Dec 04 19:28:37 crc kubenswrapper[4733]: I1204 19:28:37.334093 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" event={"ID":"fbe1ced6-e57e-4add-8c40-a56fd21e055b","Type":"ContainerStarted","Data":"0e0d1185a01a2e01e124aeadf9f46877032813ef9c6309ca4c5214bfac903308"} Dec 04 19:28:37 crc kubenswrapper[4733]: I1204 19:28:37.334921 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:37 crc kubenswrapper[4733]: I1204 19:28:37.369199 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" podStartSLOduration=3.369174352 podStartE2EDuration="3.369174352s" podCreationTimestamp="2025-12-04 19:28:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:28:37.352190331 +0000 UTC m=+6579.307551387" watchObservedRunningTime="2025-12-04 19:28:37.369174352 +0000 UTC m=+6579.324535418" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.825460 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft"] Dec 04 19:28:40 crc kubenswrapper[4733]: E1204 19:28:40.826205 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c087cdb3-7b59-4abf-be65-590d0ee4166b" containerName="init" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.826218 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c087cdb3-7b59-4abf-be65-590d0ee4166b" containerName="init" Dec 04 19:28:40 crc kubenswrapper[4733]: E1204 19:28:40.826232 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c087cdb3-7b59-4abf-be65-590d0ee4166b" containerName="dnsmasq-dns" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.826238 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c087cdb3-7b59-4abf-be65-590d0ee4166b" containerName="dnsmasq-dns" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.826421 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c087cdb3-7b59-4abf-be65-590d0ee4166b" containerName="dnsmasq-dns" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.827274 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.830173 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.830198 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.831279 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.860339 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.878865 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft"] Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.939413 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.939468 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.939633 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r8ph\" (UniqueName: \"kubernetes.io/projected/70ce3bdd-8a94-4522-a977-106089e82c98-kube-api-access-7r8ph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.939989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:40 crc kubenswrapper[4733]: I1204 19:28:40.940357 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.041958 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r8ph\" (UniqueName: \"kubernetes.io/projected/70ce3bdd-8a94-4522-a977-106089e82c98-kube-api-access-7r8ph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.042095 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.042214 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.042261 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.042493 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.049507 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.050150 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.050403 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.053711 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.067899 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r8ph\" (UniqueName: \"kubernetes.io/projected/70ce3bdd-8a94-4522-a977-106089e82c98-kube-api-access-7r8ph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c9krft\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.164811 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:28:41 crc kubenswrapper[4733]: I1204 19:28:41.755651 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft"] Dec 04 19:28:41 crc kubenswrapper[4733]: W1204 19:28:41.758078 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70ce3bdd_8a94_4522_a977_106089e82c98.slice/crio-6fa6fe914eb3c0abb9ffe98271392633b31dc680adc8948d0413be382a4d0798 WatchSource:0}: Error finding container 6fa6fe914eb3c0abb9ffe98271392633b31dc680adc8948d0413be382a4d0798: Status 404 returned error can't find the container with id 6fa6fe914eb3c0abb9ffe98271392633b31dc680adc8948d0413be382a4d0798 Dec 04 19:28:42 crc kubenswrapper[4733]: I1204 19:28:42.388062 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" event={"ID":"70ce3bdd-8a94-4522-a977-106089e82c98","Type":"ContainerStarted","Data":"6fa6fe914eb3c0abb9ffe98271392633b31dc680adc8948d0413be382a4d0798"} Dec 04 19:28:44 crc kubenswrapper[4733]: I1204 19:28:44.782942 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d959b8c7-kxj9z" Dec 04 19:28:44 crc kubenswrapper[4733]: I1204 19:28:44.877723 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9cfcfbc7-xh59n"] Dec 04 19:28:44 crc kubenswrapper[4733]: I1204 19:28:44.877965 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" podUID="e7463837-6e34-4089-bd43-8912ec048984" containerName="dnsmasq-dns" containerID="cri-o://c1e58f880faa02d2aa18227d05a3d2292ccc8e244b975d5a514d51904264da56" gracePeriod=10 Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.434969 4733 generic.go:334] "Generic (PLEG): container finished" podID="e7463837-6e34-4089-bd43-8912ec048984" containerID="c1e58f880faa02d2aa18227d05a3d2292ccc8e244b975d5a514d51904264da56" exitCode=0 Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.435052 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" event={"ID":"e7463837-6e34-4089-bd43-8912ec048984","Type":"ContainerDied","Data":"c1e58f880faa02d2aa18227d05a3d2292ccc8e244b975d5a514d51904264da56"} Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.435266 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" event={"ID":"e7463837-6e34-4089-bd43-8912ec048984","Type":"ContainerDied","Data":"3ec79b28d05d511cc3a1fda8d8443fd579f9f6c8d0639a7e8287d6119aeb1112"} Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.435282 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ec79b28d05d511cc3a1fda8d8443fd579f9f6c8d0639a7e8287d6119aeb1112" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.455548 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.547541 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-openstack-cell1\") pod \"e7463837-6e34-4089-bd43-8912ec048984\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.547581 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-dns-svc\") pod \"e7463837-6e34-4089-bd43-8912ec048984\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.547600 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-config\") pod \"e7463837-6e34-4089-bd43-8912ec048984\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.547626 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gjd6\" (UniqueName: \"kubernetes.io/projected/e7463837-6e34-4089-bd43-8912ec048984-kube-api-access-5gjd6\") pod \"e7463837-6e34-4089-bd43-8912ec048984\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.547706 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-nb\") pod \"e7463837-6e34-4089-bd43-8912ec048984\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.547782 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-sb\") pod \"e7463837-6e34-4089-bd43-8912ec048984\" (UID: \"e7463837-6e34-4089-bd43-8912ec048984\") " Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.578259 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7463837-6e34-4089-bd43-8912ec048984-kube-api-access-5gjd6" (OuterVolumeSpecName: "kube-api-access-5gjd6") pod "e7463837-6e34-4089-bd43-8912ec048984" (UID: "e7463837-6e34-4089-bd43-8912ec048984"). InnerVolumeSpecName "kube-api-access-5gjd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.607353 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e7463837-6e34-4089-bd43-8912ec048984" (UID: "e7463837-6e34-4089-bd43-8912ec048984"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.616535 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-config" (OuterVolumeSpecName: "config") pod "e7463837-6e34-4089-bd43-8912ec048984" (UID: "e7463837-6e34-4089-bd43-8912ec048984"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.619551 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e7463837-6e34-4089-bd43-8912ec048984" (UID: "e7463837-6e34-4089-bd43-8912ec048984"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.629469 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "e7463837-6e34-4089-bd43-8912ec048984" (UID: "e7463837-6e34-4089-bd43-8912ec048984"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.652462 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.652514 4733 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.652527 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-config\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.652537 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gjd6\" (UniqueName: \"kubernetes.io/projected/e7463837-6e34-4089-bd43-8912ec048984-kube-api-access-5gjd6\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.652547 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.666551 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e7463837-6e34-4089-bd43-8912ec048984" (UID: "e7463837-6e34-4089-bd43-8912ec048984"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:28:45 crc kubenswrapper[4733]: I1204 19:28:45.754704 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7463837-6e34-4089-bd43-8912ec048984-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 19:28:46 crc kubenswrapper[4733]: I1204 19:28:46.338285 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:28:46 crc kubenswrapper[4733]: E1204 19:28:46.338956 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:28:46 crc kubenswrapper[4733]: I1204 19:28:46.448102 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9cfcfbc7-xh59n" Dec 04 19:28:46 crc kubenswrapper[4733]: I1204 19:28:46.487895 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9cfcfbc7-xh59n"] Dec 04 19:28:46 crc kubenswrapper[4733]: I1204 19:28:46.502304 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9cfcfbc7-xh59n"] Dec 04 19:28:48 crc kubenswrapper[4733]: I1204 19:28:48.347836 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7463837-6e34-4089-bd43-8912ec048984" path="/var/lib/kubelet/pods/e7463837-6e34-4089-bd43-8912ec048984/volumes" Dec 04 19:28:52 crc kubenswrapper[4733]: I1204 19:28:52.537249 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" event={"ID":"70ce3bdd-8a94-4522-a977-106089e82c98","Type":"ContainerStarted","Data":"7747b18b8f5311761e014fb5787dd26fb06f43f7adb8b86914d9e85ebc705757"} Dec 04 19:28:52 crc kubenswrapper[4733]: I1204 19:28:52.564962 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" podStartSLOduration=2.318552844 podStartE2EDuration="12.564939136s" podCreationTimestamp="2025-12-04 19:28:40 +0000 UTC" firstStartedPulling="2025-12-04 19:28:41.760541463 +0000 UTC m=+6583.715902519" lastFinishedPulling="2025-12-04 19:28:52.006927725 +0000 UTC m=+6593.962288811" observedRunningTime="2025-12-04 19:28:52.556839607 +0000 UTC m=+6594.512200683" watchObservedRunningTime="2025-12-04 19:28:52.564939136 +0000 UTC m=+6594.520300182" Dec 04 19:28:59 crc kubenswrapper[4733]: I1204 19:28:59.335893 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:28:59 crc kubenswrapper[4733]: E1204 19:28:59.336566 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:29:05 crc kubenswrapper[4733]: I1204 19:29:05.666854 4733 generic.go:334] "Generic (PLEG): container finished" podID="70ce3bdd-8a94-4522-a977-106089e82c98" containerID="7747b18b8f5311761e014fb5787dd26fb06f43f7adb8b86914d9e85ebc705757" exitCode=0 Dec 04 19:29:05 crc kubenswrapper[4733]: I1204 19:29:05.666962 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" event={"ID":"70ce3bdd-8a94-4522-a977-106089e82c98","Type":"ContainerDied","Data":"7747b18b8f5311761e014fb5787dd26fb06f43f7adb8b86914d9e85ebc705757"} Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.158133 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.289503 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ssh-key\") pod \"70ce3bdd-8a94-4522-a977-106089e82c98\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.289549 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ceph\") pod \"70ce3bdd-8a94-4522-a977-106089e82c98\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.289658 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-inventory\") pod \"70ce3bdd-8a94-4522-a977-106089e82c98\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.289705 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-pre-adoption-validation-combined-ca-bundle\") pod \"70ce3bdd-8a94-4522-a977-106089e82c98\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.289976 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r8ph\" (UniqueName: \"kubernetes.io/projected/70ce3bdd-8a94-4522-a977-106089e82c98-kube-api-access-7r8ph\") pod \"70ce3bdd-8a94-4522-a977-106089e82c98\" (UID: \"70ce3bdd-8a94-4522-a977-106089e82c98\") " Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.299066 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70ce3bdd-8a94-4522-a977-106089e82c98-kube-api-access-7r8ph" (OuterVolumeSpecName: "kube-api-access-7r8ph") pod "70ce3bdd-8a94-4522-a977-106089e82c98" (UID: "70ce3bdd-8a94-4522-a977-106089e82c98"). InnerVolumeSpecName "kube-api-access-7r8ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.299205 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "70ce3bdd-8a94-4522-a977-106089e82c98" (UID: "70ce3bdd-8a94-4522-a977-106089e82c98"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.314697 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ceph" (OuterVolumeSpecName: "ceph") pod "70ce3bdd-8a94-4522-a977-106089e82c98" (UID: "70ce3bdd-8a94-4522-a977-106089e82c98"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.324635 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-inventory" (OuterVolumeSpecName: "inventory") pod "70ce3bdd-8a94-4522-a977-106089e82c98" (UID: "70ce3bdd-8a94-4522-a977-106089e82c98"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.341814 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "70ce3bdd-8a94-4522-a977-106089e82c98" (UID: "70ce3bdd-8a94-4522-a977-106089e82c98"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.392812 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.392848 4733 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.392863 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r8ph\" (UniqueName: \"kubernetes.io/projected/70ce3bdd-8a94-4522-a977-106089e82c98-kube-api-access-7r8ph\") on node \"crc\" DevicePath \"\"" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.392876 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.392886 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/70ce3bdd-8a94-4522-a977-106089e82c98-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.689367 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" event={"ID":"70ce3bdd-8a94-4522-a977-106089e82c98","Type":"ContainerDied","Data":"6fa6fe914eb3c0abb9ffe98271392633b31dc680adc8948d0413be382a4d0798"} Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.689416 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fa6fe914eb3c0abb9ffe98271392633b31dc680adc8948d0413be382a4d0798" Dec 04 19:29:07 crc kubenswrapper[4733]: I1204 19:29:07.689417 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c9krft" Dec 04 19:29:12 crc kubenswrapper[4733]: I1204 19:29:12.335540 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:29:12 crc kubenswrapper[4733]: E1204 19:29:12.336470 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.413201 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2"] Dec 04 19:29:18 crc kubenswrapper[4733]: E1204 19:29:18.414314 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7463837-6e34-4089-bd43-8912ec048984" containerName="init" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.414333 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7463837-6e34-4089-bd43-8912ec048984" containerName="init" Dec 04 19:29:18 crc kubenswrapper[4733]: E1204 19:29:18.414357 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7463837-6e34-4089-bd43-8912ec048984" containerName="dnsmasq-dns" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.414365 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7463837-6e34-4089-bd43-8912ec048984" containerName="dnsmasq-dns" Dec 04 19:29:18 crc kubenswrapper[4733]: E1204 19:29:18.414387 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ce3bdd-8a94-4522-a977-106089e82c98" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.414397 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ce3bdd-8a94-4522-a977-106089e82c98" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.414699 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ce3bdd-8a94-4522-a977-106089e82c98" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.414726 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7463837-6e34-4089-bd43-8912ec048984" containerName="dnsmasq-dns" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.415697 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.421396 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.421488 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.421581 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.421876 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.427451 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2"] Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.546058 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhrsw\" (UniqueName: \"kubernetes.io/projected/3dacba4e-76b0-40f9-92b8-67feb437f1a7-kube-api-access-bhrsw\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.546178 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.546266 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.546309 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.546358 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.648157 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.648242 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.648303 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.648393 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhrsw\" (UniqueName: \"kubernetes.io/projected/3dacba4e-76b0-40f9-92b8-67feb437f1a7-kube-api-access-bhrsw\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.648439 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.657915 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.669151 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhrsw\" (UniqueName: \"kubernetes.io/projected/3dacba4e-76b0-40f9-92b8-67feb437f1a7-kube-api-access-bhrsw\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.669163 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.673104 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.675951 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:18 crc kubenswrapper[4733]: I1204 19:29:18.756022 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:29:19 crc kubenswrapper[4733]: I1204 19:29:19.308157 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2"] Dec 04 19:29:19 crc kubenswrapper[4733]: I1204 19:29:19.824716 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" event={"ID":"3dacba4e-76b0-40f9-92b8-67feb437f1a7","Type":"ContainerStarted","Data":"728c39afcb66cc16ed43a83df26af29cab638cba9efcaec754a2951883c2b1ac"} Dec 04 19:29:20 crc kubenswrapper[4733]: I1204 19:29:20.832950 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" event={"ID":"3dacba4e-76b0-40f9-92b8-67feb437f1a7","Type":"ContainerStarted","Data":"c88c38519f002e57eb91ee0451a676a859188649e782903da78a89e088707da7"} Dec 04 19:29:20 crc kubenswrapper[4733]: I1204 19:29:20.854999 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" podStartSLOduration=2.380172168 podStartE2EDuration="2.854974796s" podCreationTimestamp="2025-12-04 19:29:18 +0000 UTC" firstStartedPulling="2025-12-04 19:29:19.312948177 +0000 UTC m=+6621.268309263" lastFinishedPulling="2025-12-04 19:29:19.787750845 +0000 UTC m=+6621.743111891" observedRunningTime="2025-12-04 19:29:20.847177455 +0000 UTC m=+6622.802538511" watchObservedRunningTime="2025-12-04 19:29:20.854974796 +0000 UTC m=+6622.810335842" Dec 04 19:29:24 crc kubenswrapper[4733]: I1204 19:29:24.335846 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:29:24 crc kubenswrapper[4733]: E1204 19:29:24.336670 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:29:26 crc kubenswrapper[4733]: I1204 19:29:26.191449 4733 scope.go:117] "RemoveContainer" containerID="a21d7cd81789636e821ed3a2ace578ada3f0add82ba799370a593fc028d5a66a" Dec 04 19:29:26 crc kubenswrapper[4733]: I1204 19:29:26.389950 4733 scope.go:117] "RemoveContainer" containerID="6e8e04ac655bc1e690d341e87f55fd2542bdb0cb52355f57a6d5f6a875cca5dd" Dec 04 19:29:26 crc kubenswrapper[4733]: I1204 19:29:26.408785 4733 scope.go:117] "RemoveContainer" containerID="a2983d8c75c7e3fcad745d5a5baf4a79c8dd1311551c099400fbd249fc5dea54" Dec 04 19:29:26 crc kubenswrapper[4733]: I1204 19:29:26.427950 4733 scope.go:117] "RemoveContainer" containerID="f3411c5af1bcd50e9844b7232f6d9a0fbde9b34377784d4c9df3831afffa4e99" Dec 04 19:29:38 crc kubenswrapper[4733]: I1204 19:29:38.352199 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:29:38 crc kubenswrapper[4733]: E1204 19:29:38.353545 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:29:53 crc kubenswrapper[4733]: I1204 19:29:53.044586 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-qfl9v"] Dec 04 19:29:53 crc kubenswrapper[4733]: I1204 19:29:53.056344 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-qfl9v"] Dec 04 19:29:53 crc kubenswrapper[4733]: I1204 19:29:53.336391 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:29:53 crc kubenswrapper[4733]: E1204 19:29:53.336687 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:29:54 crc kubenswrapper[4733]: I1204 19:29:54.355077 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d8dacc3-457b-4cbc-b8c8-56ac4e333e17" path="/var/lib/kubelet/pods/1d8dacc3-457b-4cbc-b8c8-56ac4e333e17/volumes" Dec 04 19:29:55 crc kubenswrapper[4733]: I1204 19:29:55.040423 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-c6b6-account-create-update-f94nx"] Dec 04 19:29:55 crc kubenswrapper[4733]: I1204 19:29:55.053718 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-c6b6-account-create-update-f94nx"] Dec 04 19:29:56 crc kubenswrapper[4733]: I1204 19:29:56.350625 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5023d9fb-76ff-4362-b564-c54737351d65" path="/var/lib/kubelet/pods/5023d9fb-76ff-4362-b564-c54737351d65/volumes" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.038730 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-6lknq"] Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.059604 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-6lknq"] Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.151423 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs"] Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.153048 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.155657 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.157980 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.160616 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs"] Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.246160 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/91fcf596-4fd0-4332-99e7-376ae3dfac29-config-volume\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.246324 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftb9g\" (UniqueName: \"kubernetes.io/projected/91fcf596-4fd0-4332-99e7-376ae3dfac29-kube-api-access-ftb9g\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.246421 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/91fcf596-4fd0-4332-99e7-376ae3dfac29-secret-volume\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.347343 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27cc81ff-912a-46e8-94ad-3a0dfd3a9085" path="/var/lib/kubelet/pods/27cc81ff-912a-46e8-94ad-3a0dfd3a9085/volumes" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.348524 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/91fcf596-4fd0-4332-99e7-376ae3dfac29-config-volume\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.348599 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftb9g\" (UniqueName: \"kubernetes.io/projected/91fcf596-4fd0-4332-99e7-376ae3dfac29-kube-api-access-ftb9g\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.348650 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/91fcf596-4fd0-4332-99e7-376ae3dfac29-secret-volume\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.350090 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/91fcf596-4fd0-4332-99e7-376ae3dfac29-config-volume\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.360454 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/91fcf596-4fd0-4332-99e7-376ae3dfac29-secret-volume\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.366631 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftb9g\" (UniqueName: \"kubernetes.io/projected/91fcf596-4fd0-4332-99e7-376ae3dfac29-kube-api-access-ftb9g\") pod \"collect-profiles-29414610-wx5vs\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.472996 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:00 crc kubenswrapper[4733]: I1204 19:30:00.921113 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs"] Dec 04 19:30:01 crc kubenswrapper[4733]: I1204 19:30:01.042894 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-7a0a-account-create-update-45wx9"] Dec 04 19:30:01 crc kubenswrapper[4733]: I1204 19:30:01.052660 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-7a0a-account-create-update-45wx9"] Dec 04 19:30:01 crc kubenswrapper[4733]: I1204 19:30:01.305481 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" event={"ID":"91fcf596-4fd0-4332-99e7-376ae3dfac29","Type":"ContainerStarted","Data":"5ad51377fd89bff828aeb2e6411eaba6e29b3e3bd1b51c2e8be3320cbeda7f88"} Dec 04 19:30:01 crc kubenswrapper[4733]: I1204 19:30:01.305539 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" event={"ID":"91fcf596-4fd0-4332-99e7-376ae3dfac29","Type":"ContainerStarted","Data":"882cd5800e2df055d2b274d233895c179a8383f2f66344f067aca79b4ff03c02"} Dec 04 19:30:01 crc kubenswrapper[4733]: I1204 19:30:01.332598 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" podStartSLOduration=1.332580839 podStartE2EDuration="1.332580839s" podCreationTimestamp="2025-12-04 19:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:30:01.328100417 +0000 UTC m=+6663.283461473" watchObservedRunningTime="2025-12-04 19:30:01.332580839 +0000 UTC m=+6663.287941885" Dec 04 19:30:02 crc kubenswrapper[4733]: I1204 19:30:02.361005 4733 generic.go:334] "Generic (PLEG): container finished" podID="91fcf596-4fd0-4332-99e7-376ae3dfac29" containerID="5ad51377fd89bff828aeb2e6411eaba6e29b3e3bd1b51c2e8be3320cbeda7f88" exitCode=0 Dec 04 19:30:02 crc kubenswrapper[4733]: I1204 19:30:02.375264 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a62c89-bc73-48b3-9bdf-775f5a3f9bdd" path="/var/lib/kubelet/pods/14a62c89-bc73-48b3-9bdf-775f5a3f9bdd/volumes" Dec 04 19:30:02 crc kubenswrapper[4733]: I1204 19:30:02.376289 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" event={"ID":"91fcf596-4fd0-4332-99e7-376ae3dfac29","Type":"ContainerDied","Data":"5ad51377fd89bff828aeb2e6411eaba6e29b3e3bd1b51c2e8be3320cbeda7f88"} Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.769894 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.820836 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftb9g\" (UniqueName: \"kubernetes.io/projected/91fcf596-4fd0-4332-99e7-376ae3dfac29-kube-api-access-ftb9g\") pod \"91fcf596-4fd0-4332-99e7-376ae3dfac29\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.821017 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/91fcf596-4fd0-4332-99e7-376ae3dfac29-config-volume\") pod \"91fcf596-4fd0-4332-99e7-376ae3dfac29\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.821081 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/91fcf596-4fd0-4332-99e7-376ae3dfac29-secret-volume\") pod \"91fcf596-4fd0-4332-99e7-376ae3dfac29\" (UID: \"91fcf596-4fd0-4332-99e7-376ae3dfac29\") " Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.821663 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91fcf596-4fd0-4332-99e7-376ae3dfac29-config-volume" (OuterVolumeSpecName: "config-volume") pod "91fcf596-4fd0-4332-99e7-376ae3dfac29" (UID: "91fcf596-4fd0-4332-99e7-376ae3dfac29"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.828551 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91fcf596-4fd0-4332-99e7-376ae3dfac29-kube-api-access-ftb9g" (OuterVolumeSpecName: "kube-api-access-ftb9g") pod "91fcf596-4fd0-4332-99e7-376ae3dfac29" (UID: "91fcf596-4fd0-4332-99e7-376ae3dfac29"). InnerVolumeSpecName "kube-api-access-ftb9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.829679 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91fcf596-4fd0-4332-99e7-376ae3dfac29-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "91fcf596-4fd0-4332-99e7-376ae3dfac29" (UID: "91fcf596-4fd0-4332-99e7-376ae3dfac29"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.924201 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/91fcf596-4fd0-4332-99e7-376ae3dfac29-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.924233 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/91fcf596-4fd0-4332-99e7-376ae3dfac29-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 19:30:03 crc kubenswrapper[4733]: I1204 19:30:03.924245 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftb9g\" (UniqueName: \"kubernetes.io/projected/91fcf596-4fd0-4332-99e7-376ae3dfac29-kube-api-access-ftb9g\") on node \"crc\" DevicePath \"\"" Dec 04 19:30:04 crc kubenswrapper[4733]: I1204 19:30:04.335062 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:30:04 crc kubenswrapper[4733]: E1204 19:30:04.335518 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:30:04 crc kubenswrapper[4733]: I1204 19:30:04.385101 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" event={"ID":"91fcf596-4fd0-4332-99e7-376ae3dfac29","Type":"ContainerDied","Data":"882cd5800e2df055d2b274d233895c179a8383f2f66344f067aca79b4ff03c02"} Dec 04 19:30:04 crc kubenswrapper[4733]: I1204 19:30:04.385176 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="882cd5800e2df055d2b274d233895c179a8383f2f66344f067aca79b4ff03c02" Dec 04 19:30:04 crc kubenswrapper[4733]: I1204 19:30:04.385184 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs" Dec 04 19:30:04 crc kubenswrapper[4733]: I1204 19:30:04.424555 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv"] Dec 04 19:30:04 crc kubenswrapper[4733]: I1204 19:30:04.457684 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414565-kd7vv"] Dec 04 19:30:06 crc kubenswrapper[4733]: I1204 19:30:06.361176 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bfcbd43-0013-4715-81ad-dd378598b7e1" path="/var/lib/kubelet/pods/7bfcbd43-0013-4715-81ad-dd378598b7e1/volumes" Dec 04 19:30:19 crc kubenswrapper[4733]: I1204 19:30:19.336703 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:30:19 crc kubenswrapper[4733]: E1204 19:30:19.337840 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:30:25 crc kubenswrapper[4733]: I1204 19:30:25.964211 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xxd9c"] Dec 04 19:30:25 crc kubenswrapper[4733]: E1204 19:30:25.965257 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fcf596-4fd0-4332-99e7-376ae3dfac29" containerName="collect-profiles" Dec 04 19:30:25 crc kubenswrapper[4733]: I1204 19:30:25.965280 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fcf596-4fd0-4332-99e7-376ae3dfac29" containerName="collect-profiles" Dec 04 19:30:25 crc kubenswrapper[4733]: I1204 19:30:25.965699 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="91fcf596-4fd0-4332-99e7-376ae3dfac29" containerName="collect-profiles" Dec 04 19:30:25 crc kubenswrapper[4733]: I1204 19:30:25.968081 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:25 crc kubenswrapper[4733]: I1204 19:30:25.978026 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xxd9c"] Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.020878 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84xkn\" (UniqueName: \"kubernetes.io/projected/e10c89cf-8a38-4a98-8400-0832c3788c94-kube-api-access-84xkn\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.020934 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-utilities\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.021218 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-catalog-content\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.123157 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-utilities\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.123519 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-catalog-content\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.123658 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84xkn\" (UniqueName: \"kubernetes.io/projected/e10c89cf-8a38-4a98-8400-0832c3788c94-kube-api-access-84xkn\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.123714 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-utilities\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.124027 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-catalog-content\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.143728 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84xkn\" (UniqueName: \"kubernetes.io/projected/e10c89cf-8a38-4a98-8400-0832c3788c94-kube-api-access-84xkn\") pod \"certified-operators-xxd9c\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.290919 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.677853 4733 scope.go:117] "RemoveContainer" containerID="af7127c73e9826cf31c7d42cd1e66a5bdbccaf09d08eba57f9d1df90daee5456" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.703511 4733 scope.go:117] "RemoveContainer" containerID="9b1e2cbd8b82c3ecda12f0035b6050e8539ff39af8334049c19df65422266c2e" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.749699 4733 scope.go:117] "RemoveContainer" containerID="80a849d764c8dbfb9752cfdef552031d425786c4297340190f3d973b5de8efe2" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.765526 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xxd9c"] Dec 04 19:30:26 crc kubenswrapper[4733]: W1204 19:30:26.774057 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode10c89cf_8a38_4a98_8400_0832c3788c94.slice/crio-c7c213d6910319d46076df511fd288c0aa43f63d03492ae208f72bfb74bcfd07 WatchSource:0}: Error finding container c7c213d6910319d46076df511fd288c0aa43f63d03492ae208f72bfb74bcfd07: Status 404 returned error can't find the container with id c7c213d6910319d46076df511fd288c0aa43f63d03492ae208f72bfb74bcfd07 Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.829967 4733 scope.go:117] "RemoveContainer" containerID="68f2489219caec47d0bf068e927c8f7687dd938ce8b6887f37c6f1e139abeb85" Dec 04 19:30:26 crc kubenswrapper[4733]: I1204 19:30:26.858143 4733 scope.go:117] "RemoveContainer" containerID="6a500f5d0dce38480ed2ff82385fed20f4fabffc2e397f8601d3ec631f310d02" Dec 04 19:30:27 crc kubenswrapper[4733]: I1204 19:30:27.668111 4733 generic.go:334] "Generic (PLEG): container finished" podID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerID="1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9" exitCode=0 Dec 04 19:30:27 crc kubenswrapper[4733]: I1204 19:30:27.668178 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxd9c" event={"ID":"e10c89cf-8a38-4a98-8400-0832c3788c94","Type":"ContainerDied","Data":"1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9"} Dec 04 19:30:27 crc kubenswrapper[4733]: I1204 19:30:27.668390 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxd9c" event={"ID":"e10c89cf-8a38-4a98-8400-0832c3788c94","Type":"ContainerStarted","Data":"c7c213d6910319d46076df511fd288c0aa43f63d03492ae208f72bfb74bcfd07"} Dec 04 19:30:27 crc kubenswrapper[4733]: I1204 19:30:27.670419 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:30:29 crc kubenswrapper[4733]: I1204 19:30:29.694144 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxd9c" event={"ID":"e10c89cf-8a38-4a98-8400-0832c3788c94","Type":"ContainerStarted","Data":"72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8"} Dec 04 19:30:30 crc kubenswrapper[4733]: I1204 19:30:30.708332 4733 generic.go:334] "Generic (PLEG): container finished" podID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerID="72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8" exitCode=0 Dec 04 19:30:30 crc kubenswrapper[4733]: I1204 19:30:30.708408 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxd9c" event={"ID":"e10c89cf-8a38-4a98-8400-0832c3788c94","Type":"ContainerDied","Data":"72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8"} Dec 04 19:30:31 crc kubenswrapper[4733]: I1204 19:30:31.723146 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxd9c" event={"ID":"e10c89cf-8a38-4a98-8400-0832c3788c94","Type":"ContainerStarted","Data":"9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96"} Dec 04 19:30:31 crc kubenswrapper[4733]: I1204 19:30:31.747576 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xxd9c" podStartSLOduration=3.315736171 podStartE2EDuration="6.747557253s" podCreationTimestamp="2025-12-04 19:30:25 +0000 UTC" firstStartedPulling="2025-12-04 19:30:27.67011541 +0000 UTC m=+6689.625476466" lastFinishedPulling="2025-12-04 19:30:31.101936502 +0000 UTC m=+6693.057297548" observedRunningTime="2025-12-04 19:30:31.742343271 +0000 UTC m=+6693.697704337" watchObservedRunningTime="2025-12-04 19:30:31.747557253 +0000 UTC m=+6693.702918299" Dec 04 19:30:34 crc kubenswrapper[4733]: I1204 19:30:34.336144 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:30:34 crc kubenswrapper[4733]: E1204 19:30:34.336876 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:30:36 crc kubenswrapper[4733]: I1204 19:30:36.291189 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:36 crc kubenswrapper[4733]: I1204 19:30:36.291661 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:36 crc kubenswrapper[4733]: I1204 19:30:36.333615 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:36 crc kubenswrapper[4733]: I1204 19:30:36.841934 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:36 crc kubenswrapper[4733]: I1204 19:30:36.895931 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xxd9c"] Dec 04 19:30:38 crc kubenswrapper[4733]: I1204 19:30:38.801833 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xxd9c" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerName="registry-server" containerID="cri-o://9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96" gracePeriod=2 Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.296999 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.432441 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84xkn\" (UniqueName: \"kubernetes.io/projected/e10c89cf-8a38-4a98-8400-0832c3788c94-kube-api-access-84xkn\") pod \"e10c89cf-8a38-4a98-8400-0832c3788c94\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.432586 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-catalog-content\") pod \"e10c89cf-8a38-4a98-8400-0832c3788c94\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.432769 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-utilities\") pod \"e10c89cf-8a38-4a98-8400-0832c3788c94\" (UID: \"e10c89cf-8a38-4a98-8400-0832c3788c94\") " Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.435236 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-utilities" (OuterVolumeSpecName: "utilities") pod "e10c89cf-8a38-4a98-8400-0832c3788c94" (UID: "e10c89cf-8a38-4a98-8400-0832c3788c94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.439139 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e10c89cf-8a38-4a98-8400-0832c3788c94-kube-api-access-84xkn" (OuterVolumeSpecName: "kube-api-access-84xkn") pod "e10c89cf-8a38-4a98-8400-0832c3788c94" (UID: "e10c89cf-8a38-4a98-8400-0832c3788c94"). InnerVolumeSpecName "kube-api-access-84xkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.490925 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e10c89cf-8a38-4a98-8400-0832c3788c94" (UID: "e10c89cf-8a38-4a98-8400-0832c3788c94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.535150 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.535394 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84xkn\" (UniqueName: \"kubernetes.io/projected/e10c89cf-8a38-4a98-8400-0832c3788c94-kube-api-access-84xkn\") on node \"crc\" DevicePath \"\"" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.535463 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10c89cf-8a38-4a98-8400-0832c3788c94-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.814562 4733 generic.go:334] "Generic (PLEG): container finished" podID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerID="9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96" exitCode=0 Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.814633 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxd9c" event={"ID":"e10c89cf-8a38-4a98-8400-0832c3788c94","Type":"ContainerDied","Data":"9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96"} Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.814827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxd9c" event={"ID":"e10c89cf-8a38-4a98-8400-0832c3788c94","Type":"ContainerDied","Data":"c7c213d6910319d46076df511fd288c0aa43f63d03492ae208f72bfb74bcfd07"} Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.814845 4733 scope.go:117] "RemoveContainer" containerID="9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.814701 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxd9c" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.836825 4733 scope.go:117] "RemoveContainer" containerID="72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.862620 4733 scope.go:117] "RemoveContainer" containerID="1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.874110 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xxd9c"] Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.883210 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xxd9c"] Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.916567 4733 scope.go:117] "RemoveContainer" containerID="9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96" Dec 04 19:30:39 crc kubenswrapper[4733]: E1204 19:30:39.916958 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96\": container with ID starting with 9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96 not found: ID does not exist" containerID="9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.916994 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96"} err="failed to get container status \"9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96\": rpc error: code = NotFound desc = could not find container \"9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96\": container with ID starting with 9a2c90ef48de1d53fa3c5f9c3020c443a34247c67d0bf8736e727b987f0aab96 not found: ID does not exist" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.917021 4733 scope.go:117] "RemoveContainer" containerID="72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8" Dec 04 19:30:39 crc kubenswrapper[4733]: E1204 19:30:39.917411 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8\": container with ID starting with 72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8 not found: ID does not exist" containerID="72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.917434 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8"} err="failed to get container status \"72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8\": rpc error: code = NotFound desc = could not find container \"72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8\": container with ID starting with 72a52be71fa4e75cbe5df9698d94b2bfc7a39deec06fdae0cc02bd11e82793a8 not found: ID does not exist" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.917465 4733 scope.go:117] "RemoveContainer" containerID="1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9" Dec 04 19:30:39 crc kubenswrapper[4733]: E1204 19:30:39.917822 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9\": container with ID starting with 1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9 not found: ID does not exist" containerID="1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9" Dec 04 19:30:39 crc kubenswrapper[4733]: I1204 19:30:39.917887 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9"} err="failed to get container status \"1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9\": rpc error: code = NotFound desc = could not find container \"1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9\": container with ID starting with 1a8309a3feec8c75a5fa57d2231fc191d0130ac5107a9703ef3e402b728772d9 not found: ID does not exist" Dec 04 19:30:40 crc kubenswrapper[4733]: I1204 19:30:40.357901 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" path="/var/lib/kubelet/pods/e10c89cf-8a38-4a98-8400-0832c3788c94/volumes" Dec 04 19:30:43 crc kubenswrapper[4733]: I1204 19:30:43.046465 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-r2p6q"] Dec 04 19:30:43 crc kubenswrapper[4733]: I1204 19:30:43.058099 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-r2p6q"] Dec 04 19:30:44 crc kubenswrapper[4733]: I1204 19:30:44.351730 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f89bf105-5dad-4c4e-a810-d9cc9936cf48" path="/var/lib/kubelet/pods/f89bf105-5dad-4c4e-a810-d9cc9936cf48/volumes" Dec 04 19:30:46 crc kubenswrapper[4733]: I1204 19:30:46.335957 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:30:46 crc kubenswrapper[4733]: E1204 19:30:46.336503 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:30:59 crc kubenswrapper[4733]: I1204 19:30:59.336563 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:30:59 crc kubenswrapper[4733]: E1204 19:30:59.337544 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:31:14 crc kubenswrapper[4733]: I1204 19:31:14.336061 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:31:14 crc kubenswrapper[4733]: E1204 19:31:14.337062 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:31:26 crc kubenswrapper[4733]: I1204 19:31:26.987027 4733 scope.go:117] "RemoveContainer" containerID="95b64fb5660e9d6af7f652b88cd625bee0b60b8376b5f9d07e7db25bf6eca53d" Dec 04 19:31:27 crc kubenswrapper[4733]: I1204 19:31:27.028332 4733 scope.go:117] "RemoveContainer" containerID="6dc9a47033c60a204919e3488932469af34f67781518ae5452ec27729bfc6200" Dec 04 19:31:29 crc kubenswrapper[4733]: I1204 19:31:29.335943 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:31:29 crc kubenswrapper[4733]: E1204 19:31:29.336908 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:31:43 crc kubenswrapper[4733]: I1204 19:31:43.335877 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:31:43 crc kubenswrapper[4733]: E1204 19:31:43.336823 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:31:55 crc kubenswrapper[4733]: I1204 19:31:55.337041 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:31:55 crc kubenswrapper[4733]: I1204 19:31:55.670243 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"1059a4cf6d36f458e18b39d02074ccfc6d3afb83e3b089fc9d8f384eed628898"} Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.324574 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gtg25"] Dec 04 19:32:03 crc kubenswrapper[4733]: E1204 19:32:03.333354 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerName="extract-utilities" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.333378 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerName="extract-utilities" Dec 04 19:32:03 crc kubenswrapper[4733]: E1204 19:32:03.333419 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerName="registry-server" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.333428 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerName="registry-server" Dec 04 19:32:03 crc kubenswrapper[4733]: E1204 19:32:03.333443 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerName="extract-content" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.333452 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerName="extract-content" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.333699 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10c89cf-8a38-4a98-8400-0832c3788c94" containerName="registry-server" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.335937 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.348858 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gtg25"] Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.422251 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-catalog-content\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.422362 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6j42\" (UniqueName: \"kubernetes.io/projected/e540bc4c-1d50-4d7c-b39c-afa49d731696-kube-api-access-q6j42\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.422500 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-utilities\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.525425 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-utilities\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.525842 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-catalog-content\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.525959 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-utilities\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.526001 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6j42\" (UniqueName: \"kubernetes.io/projected/e540bc4c-1d50-4d7c-b39c-afa49d731696-kube-api-access-q6j42\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.526070 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-catalog-content\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.547199 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6j42\" (UniqueName: \"kubernetes.io/projected/e540bc4c-1d50-4d7c-b39c-afa49d731696-kube-api-access-q6j42\") pod \"redhat-operators-gtg25\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:03 crc kubenswrapper[4733]: I1204 19:32:03.669102 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:04 crc kubenswrapper[4733]: I1204 19:32:04.189540 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gtg25"] Dec 04 19:32:04 crc kubenswrapper[4733]: I1204 19:32:04.851766 4733 generic.go:334] "Generic (PLEG): container finished" podID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerID="4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4" exitCode=0 Dec 04 19:32:04 crc kubenswrapper[4733]: I1204 19:32:04.851848 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtg25" event={"ID":"e540bc4c-1d50-4d7c-b39c-afa49d731696","Type":"ContainerDied","Data":"4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4"} Dec 04 19:32:04 crc kubenswrapper[4733]: I1204 19:32:04.852107 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtg25" event={"ID":"e540bc4c-1d50-4d7c-b39c-afa49d731696","Type":"ContainerStarted","Data":"f06f37e0aba25d178779b3613914db07004d5fe96085f13e13f66dbd772d5920"} Dec 04 19:32:05 crc kubenswrapper[4733]: I1204 19:32:05.863235 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtg25" event={"ID":"e540bc4c-1d50-4d7c-b39c-afa49d731696","Type":"ContainerStarted","Data":"158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19"} Dec 04 19:32:08 crc kubenswrapper[4733]: I1204 19:32:08.896839 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtg25" event={"ID":"e540bc4c-1d50-4d7c-b39c-afa49d731696","Type":"ContainerDied","Data":"158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19"} Dec 04 19:32:08 crc kubenswrapper[4733]: I1204 19:32:08.896776 4733 generic.go:334] "Generic (PLEG): container finished" podID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerID="158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19" exitCode=0 Dec 04 19:32:09 crc kubenswrapper[4733]: I1204 19:32:09.914309 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtg25" event={"ID":"e540bc4c-1d50-4d7c-b39c-afa49d731696","Type":"ContainerStarted","Data":"d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42"} Dec 04 19:32:09 crc kubenswrapper[4733]: I1204 19:32:09.947524 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gtg25" podStartSLOduration=2.473806711 podStartE2EDuration="6.947497258s" podCreationTimestamp="2025-12-04 19:32:03 +0000 UTC" firstStartedPulling="2025-12-04 19:32:04.854082142 +0000 UTC m=+6786.809443188" lastFinishedPulling="2025-12-04 19:32:09.327772669 +0000 UTC m=+6791.283133735" observedRunningTime="2025-12-04 19:32:09.933647583 +0000 UTC m=+6791.889008629" watchObservedRunningTime="2025-12-04 19:32:09.947497258 +0000 UTC m=+6791.902858344" Dec 04 19:32:13 crc kubenswrapper[4733]: I1204 19:32:13.669829 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:13 crc kubenswrapper[4733]: I1204 19:32:13.670362 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:14 crc kubenswrapper[4733]: I1204 19:32:14.722946 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gtg25" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="registry-server" probeResult="failure" output=< Dec 04 19:32:14 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 19:32:14 crc kubenswrapper[4733]: > Dec 04 19:32:23 crc kubenswrapper[4733]: I1204 19:32:23.739864 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:23 crc kubenswrapper[4733]: I1204 19:32:23.810667 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:23 crc kubenswrapper[4733]: I1204 19:32:23.980084 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gtg25"] Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.090269 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gtg25" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="registry-server" containerID="cri-o://d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42" gracePeriod=2 Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.633321 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.723410 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-utilities\") pod \"e540bc4c-1d50-4d7c-b39c-afa49d731696\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.723632 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6j42\" (UniqueName: \"kubernetes.io/projected/e540bc4c-1d50-4d7c-b39c-afa49d731696-kube-api-access-q6j42\") pod \"e540bc4c-1d50-4d7c-b39c-afa49d731696\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.723653 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-catalog-content\") pod \"e540bc4c-1d50-4d7c-b39c-afa49d731696\" (UID: \"e540bc4c-1d50-4d7c-b39c-afa49d731696\") " Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.724207 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-utilities" (OuterVolumeSpecName: "utilities") pod "e540bc4c-1d50-4d7c-b39c-afa49d731696" (UID: "e540bc4c-1d50-4d7c-b39c-afa49d731696"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.729177 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e540bc4c-1d50-4d7c-b39c-afa49d731696-kube-api-access-q6j42" (OuterVolumeSpecName: "kube-api-access-q6j42") pod "e540bc4c-1d50-4d7c-b39c-afa49d731696" (UID: "e540bc4c-1d50-4d7c-b39c-afa49d731696"). InnerVolumeSpecName "kube-api-access-q6j42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.826196 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.826475 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6j42\" (UniqueName: \"kubernetes.io/projected/e540bc4c-1d50-4d7c-b39c-afa49d731696-kube-api-access-q6j42\") on node \"crc\" DevicePath \"\"" Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.842841 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e540bc4c-1d50-4d7c-b39c-afa49d731696" (UID: "e540bc4c-1d50-4d7c-b39c-afa49d731696"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:32:25 crc kubenswrapper[4733]: I1204 19:32:25.928703 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e540bc4c-1d50-4d7c-b39c-afa49d731696-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.104357 4733 generic.go:334] "Generic (PLEG): container finished" podID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerID="d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42" exitCode=0 Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.104395 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtg25" event={"ID":"e540bc4c-1d50-4d7c-b39c-afa49d731696","Type":"ContainerDied","Data":"d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42"} Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.104422 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtg25" event={"ID":"e540bc4c-1d50-4d7c-b39c-afa49d731696","Type":"ContainerDied","Data":"f06f37e0aba25d178779b3613914db07004d5fe96085f13e13f66dbd772d5920"} Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.104441 4733 scope.go:117] "RemoveContainer" containerID="d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.104439 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtg25" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.140557 4733 scope.go:117] "RemoveContainer" containerID="158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.142763 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gtg25"] Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.152690 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gtg25"] Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.165303 4733 scope.go:117] "RemoveContainer" containerID="4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.215140 4733 scope.go:117] "RemoveContainer" containerID="d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42" Dec 04 19:32:26 crc kubenswrapper[4733]: E1204 19:32:26.217307 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42\": container with ID starting with d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42 not found: ID does not exist" containerID="d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.217363 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42"} err="failed to get container status \"d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42\": rpc error: code = NotFound desc = could not find container \"d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42\": container with ID starting with d319dee17836a81babc29b75cc8ffd61951a3a3cb1f9501c764229f13b756f42 not found: ID does not exist" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.217401 4733 scope.go:117] "RemoveContainer" containerID="158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19" Dec 04 19:32:26 crc kubenswrapper[4733]: E1204 19:32:26.217791 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19\": container with ID starting with 158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19 not found: ID does not exist" containerID="158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.217874 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19"} err="failed to get container status \"158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19\": rpc error: code = NotFound desc = could not find container \"158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19\": container with ID starting with 158f3e5c5fecf9e18ea364607950c454f774bf7d48be488e8e06443a7ae3ed19 not found: ID does not exist" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.217917 4733 scope.go:117] "RemoveContainer" containerID="4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4" Dec 04 19:32:26 crc kubenswrapper[4733]: E1204 19:32:26.218357 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4\": container with ID starting with 4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4 not found: ID does not exist" containerID="4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.218389 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4"} err="failed to get container status \"4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4\": rpc error: code = NotFound desc = could not find container \"4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4\": container with ID starting with 4806aec6da7289bf5f481708e52bacc3709bb8b21922d732b3347e41d4c616a4 not found: ID does not exist" Dec 04 19:32:26 crc kubenswrapper[4733]: I1204 19:32:26.347215 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" path="/var/lib/kubelet/pods/e540bc4c-1d50-4d7c-b39c-afa49d731696/volumes" Dec 04 19:34:07 crc kubenswrapper[4733]: I1204 19:34:07.050865 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-ft7hm"] Dec 04 19:34:07 crc kubenswrapper[4733]: I1204 19:34:07.064000 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-ft7hm"] Dec 04 19:34:08 crc kubenswrapper[4733]: I1204 19:34:08.036940 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-2e40-account-create-update-jtgj4"] Dec 04 19:34:08 crc kubenswrapper[4733]: I1204 19:34:08.050764 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-2e40-account-create-update-jtgj4"] Dec 04 19:34:08 crc kubenswrapper[4733]: I1204 19:34:08.349561 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed7dab29-a923-4ace-811a-cad3279c16ed" path="/var/lib/kubelet/pods/ed7dab29-a923-4ace-811a-cad3279c16ed/volumes" Dec 04 19:34:08 crc kubenswrapper[4733]: I1204 19:34:08.350242 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef49d872-16fc-4e04-b146-5a7c9b260111" path="/var/lib/kubelet/pods/ef49d872-16fc-4e04-b146-5a7c9b260111/volumes" Dec 04 19:34:15 crc kubenswrapper[4733]: I1204 19:34:15.361588 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:34:15 crc kubenswrapper[4733]: I1204 19:34:15.362162 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:34:21 crc kubenswrapper[4733]: I1204 19:34:21.054332 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-xvqzk"] Dec 04 19:34:21 crc kubenswrapper[4733]: I1204 19:34:21.074700 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-xvqzk"] Dec 04 19:34:22 crc kubenswrapper[4733]: I1204 19:34:22.358239 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f7ba50-f386-4612-a44e-ae29091ce6d0" path="/var/lib/kubelet/pods/73f7ba50-f386-4612-a44e-ae29091ce6d0/volumes" Dec 04 19:34:27 crc kubenswrapper[4733]: I1204 19:34:27.214616 4733 scope.go:117] "RemoveContainer" containerID="8f9a3c448cf77fa9c7136d0326b7da058edfb480ff4aae96e94b30255ad5d50f" Dec 04 19:34:27 crc kubenswrapper[4733]: I1204 19:34:27.263075 4733 scope.go:117] "RemoveContainer" containerID="79d76ad7051f1d82b322e12e457c1bfe302b4dd31065fc2ef32982d15f5ed37e" Dec 04 19:34:27 crc kubenswrapper[4733]: I1204 19:34:27.365196 4733 scope.go:117] "RemoveContainer" containerID="c1e58f880faa02d2aa18227d05a3d2292ccc8e244b975d5a514d51904264da56" Dec 04 19:34:27 crc kubenswrapper[4733]: I1204 19:34:27.404993 4733 scope.go:117] "RemoveContainer" containerID="b88862e5b704236f458bfb9308dd8cf17815aac1abb731dc679367ed601fe860" Dec 04 19:34:27 crc kubenswrapper[4733]: I1204 19:34:27.424775 4733 scope.go:117] "RemoveContainer" containerID="b85cec2af382613c0ea142e1e44a0ad3fea9714de79af19623cd391e57505a4d" Dec 04 19:34:45 crc kubenswrapper[4733]: I1204 19:34:45.361515 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:34:45 crc kubenswrapper[4733]: I1204 19:34:45.362090 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.361686 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.362302 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.362354 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.363058 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1059a4cf6d36f458e18b39d02074ccfc6d3afb83e3b089fc9d8f384eed628898"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.363126 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://1059a4cf6d36f458e18b39d02074ccfc6d3afb83e3b089fc9d8f384eed628898" gracePeriod=600 Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.885195 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="1059a4cf6d36f458e18b39d02074ccfc6d3afb83e3b089fc9d8f384eed628898" exitCode=0 Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.885573 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"1059a4cf6d36f458e18b39d02074ccfc6d3afb83e3b089fc9d8f384eed628898"} Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.885599 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96"} Dec 04 19:35:15 crc kubenswrapper[4733]: I1204 19:35:15.885615 4733 scope.go:117] "RemoveContainer" containerID="86b5ba5a3db229c07384ba1bb91389d9d6328f5dbdb4059601268f427939ed08" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.242516 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qnxbb"] Dec 04 19:35:33 crc kubenswrapper[4733]: E1204 19:35:33.243507 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="registry-server" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.243522 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="registry-server" Dec 04 19:35:33 crc kubenswrapper[4733]: E1204 19:35:33.243539 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="extract-content" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.243547 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="extract-content" Dec 04 19:35:33 crc kubenswrapper[4733]: E1204 19:35:33.243578 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="extract-utilities" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.243586 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="extract-utilities" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.243886 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e540bc4c-1d50-4d7c-b39c-afa49d731696" containerName="registry-server" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.245715 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.267304 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnxbb"] Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.299128 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-utilities\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.299434 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-catalog-content\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.309459 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj6h8\" (UniqueName: \"kubernetes.io/projected/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-kube-api-access-gj6h8\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.411814 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj6h8\" (UniqueName: \"kubernetes.io/projected/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-kube-api-access-gj6h8\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.411963 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-utilities\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.412005 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-catalog-content\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.412591 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-catalog-content\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.412855 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-utilities\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.442359 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj6h8\" (UniqueName: \"kubernetes.io/projected/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-kube-api-access-gj6h8\") pod \"redhat-marketplace-qnxbb\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:33 crc kubenswrapper[4733]: I1204 19:35:33.577619 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:34 crc kubenswrapper[4733]: I1204 19:35:34.113204 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnxbb"] Dec 04 19:35:35 crc kubenswrapper[4733]: I1204 19:35:35.118537 4733 generic.go:334] "Generic (PLEG): container finished" podID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerID="2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc" exitCode=0 Dec 04 19:35:35 crc kubenswrapper[4733]: I1204 19:35:35.118609 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnxbb" event={"ID":"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7","Type":"ContainerDied","Data":"2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc"} Dec 04 19:35:35 crc kubenswrapper[4733]: I1204 19:35:35.118945 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnxbb" event={"ID":"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7","Type":"ContainerStarted","Data":"1527bd554d09062dd58dce0ecb725e6e86fd03c81ccf9150fdb6b7d58503bc57"} Dec 04 19:35:35 crc kubenswrapper[4733]: I1204 19:35:35.121116 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:35:37 crc kubenswrapper[4733]: I1204 19:35:37.142154 4733 generic.go:334] "Generic (PLEG): container finished" podID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerID="b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659" exitCode=0 Dec 04 19:35:37 crc kubenswrapper[4733]: I1204 19:35:37.142304 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnxbb" event={"ID":"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7","Type":"ContainerDied","Data":"b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659"} Dec 04 19:35:38 crc kubenswrapper[4733]: I1204 19:35:38.155350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnxbb" event={"ID":"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7","Type":"ContainerStarted","Data":"0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4"} Dec 04 19:35:43 crc kubenswrapper[4733]: I1204 19:35:43.588225 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:43 crc kubenswrapper[4733]: I1204 19:35:43.589052 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:43 crc kubenswrapper[4733]: I1204 19:35:43.661167 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:43 crc kubenswrapper[4733]: I1204 19:35:43.694477 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qnxbb" podStartSLOduration=8.256987592 podStartE2EDuration="10.694447045s" podCreationTimestamp="2025-12-04 19:35:33 +0000 UTC" firstStartedPulling="2025-12-04 19:35:35.120740485 +0000 UTC m=+6997.076101551" lastFinishedPulling="2025-12-04 19:35:37.558199928 +0000 UTC m=+6999.513561004" observedRunningTime="2025-12-04 19:35:38.18319344 +0000 UTC m=+7000.138554486" watchObservedRunningTime="2025-12-04 19:35:43.694447045 +0000 UTC m=+7005.649808101" Dec 04 19:35:44 crc kubenswrapper[4733]: I1204 19:35:44.280697 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:44 crc kubenswrapper[4733]: I1204 19:35:44.351084 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnxbb"] Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.239200 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qnxbb" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerName="registry-server" containerID="cri-o://0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4" gracePeriod=2 Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.766644 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.862942 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-catalog-content\") pod \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.863178 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj6h8\" (UniqueName: \"kubernetes.io/projected/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-kube-api-access-gj6h8\") pod \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.863272 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-utilities\") pod \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\" (UID: \"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7\") " Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.864075 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-utilities" (OuterVolumeSpecName: "utilities") pod "0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" (UID: "0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.868143 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-kube-api-access-gj6h8" (OuterVolumeSpecName: "kube-api-access-gj6h8") pod "0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" (UID: "0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7"). InnerVolumeSpecName "kube-api-access-gj6h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.881302 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" (UID: "0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.966011 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.966057 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj6h8\" (UniqueName: \"kubernetes.io/projected/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-kube-api-access-gj6h8\") on node \"crc\" DevicePath \"\"" Dec 04 19:35:46 crc kubenswrapper[4733]: I1204 19:35:46.966071 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.257365 4733 generic.go:334] "Generic (PLEG): container finished" podID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerID="0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4" exitCode=0 Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.257433 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnxbb" event={"ID":"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7","Type":"ContainerDied","Data":"0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4"} Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.257469 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnxbb" event={"ID":"0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7","Type":"ContainerDied","Data":"1527bd554d09062dd58dce0ecb725e6e86fd03c81ccf9150fdb6b7d58503bc57"} Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.257514 4733 scope.go:117] "RemoveContainer" containerID="0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.257834 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qnxbb" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.287142 4733 scope.go:117] "RemoveContainer" containerID="b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.309877 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnxbb"] Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.319263 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnxbb"] Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.322744 4733 scope.go:117] "RemoveContainer" containerID="2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.386146 4733 scope.go:117] "RemoveContainer" containerID="0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4" Dec 04 19:35:47 crc kubenswrapper[4733]: E1204 19:35:47.386607 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4\": container with ID starting with 0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4 not found: ID does not exist" containerID="0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.386645 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4"} err="failed to get container status \"0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4\": rpc error: code = NotFound desc = could not find container \"0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4\": container with ID starting with 0872329c4dc5962522de9c61ed4f2021927f5738ee8292cd88d75c7eff90ebf4 not found: ID does not exist" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.386668 4733 scope.go:117] "RemoveContainer" containerID="b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659" Dec 04 19:35:47 crc kubenswrapper[4733]: E1204 19:35:47.387007 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659\": container with ID starting with b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659 not found: ID does not exist" containerID="b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.387042 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659"} err="failed to get container status \"b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659\": rpc error: code = NotFound desc = could not find container \"b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659\": container with ID starting with b2f01c35e3fb3d2a7bb5dcf39984d08f7f6246691e85fe257d419c390af5c659 not found: ID does not exist" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.387061 4733 scope.go:117] "RemoveContainer" containerID="2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc" Dec 04 19:35:47 crc kubenswrapper[4733]: E1204 19:35:47.387263 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc\": container with ID starting with 2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc not found: ID does not exist" containerID="2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc" Dec 04 19:35:47 crc kubenswrapper[4733]: I1204 19:35:47.387291 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc"} err="failed to get container status \"2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc\": rpc error: code = NotFound desc = could not find container \"2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc\": container with ID starting with 2acb42b8582bf1c69b5c782f963d73849b2e60db308479c99a3c424b94a55efc not found: ID does not exist" Dec 04 19:35:48 crc kubenswrapper[4733]: I1204 19:35:48.358528 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" path="/var/lib/kubelet/pods/0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7/volumes" Dec 04 19:36:30 crc kubenswrapper[4733]: I1204 19:36:30.051988 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-5bf5z"] Dec 04 19:36:30 crc kubenswrapper[4733]: I1204 19:36:30.064116 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-af12-account-create-update-tk5q7"] Dec 04 19:36:30 crc kubenswrapper[4733]: I1204 19:36:30.073206 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-af12-account-create-update-tk5q7"] Dec 04 19:36:30 crc kubenswrapper[4733]: I1204 19:36:30.082019 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-5bf5z"] Dec 04 19:36:30 crc kubenswrapper[4733]: I1204 19:36:30.350255 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="449bb25e-8f81-4a6b-8b2a-dcca30e4f880" path="/var/lib/kubelet/pods/449bb25e-8f81-4a6b-8b2a-dcca30e4f880/volumes" Dec 04 19:36:30 crc kubenswrapper[4733]: I1204 19:36:30.351344 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcc29a4f-6e38-49ec-a8be-4464c076d391" path="/var/lib/kubelet/pods/dcc29a4f-6e38-49ec-a8be-4464c076d391/volumes" Dec 04 19:36:41 crc kubenswrapper[4733]: I1204 19:36:41.068720 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-7jggv"] Dec 04 19:36:41 crc kubenswrapper[4733]: I1204 19:36:41.080316 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-7jggv"] Dec 04 19:36:42 crc kubenswrapper[4733]: I1204 19:36:42.347696 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9595436c-1e86-4790-8ca4-3707f0d22b0c" path="/var/lib/kubelet/pods/9595436c-1e86-4790-8ca4-3707f0d22b0c/volumes" Dec 04 19:37:02 crc kubenswrapper[4733]: I1204 19:37:02.034143 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-jqdr9"] Dec 04 19:37:02 crc kubenswrapper[4733]: I1204 19:37:02.046815 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-jqdr9"] Dec 04 19:37:02 crc kubenswrapper[4733]: I1204 19:37:02.378626 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="529f9e6e-31b1-4901-90c2-47637c26fcc3" path="/var/lib/kubelet/pods/529f9e6e-31b1-4901-90c2-47637c26fcc3/volumes" Dec 04 19:37:03 crc kubenswrapper[4733]: I1204 19:37:03.033179 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-a076-account-create-update-lsnb7"] Dec 04 19:37:03 crc kubenswrapper[4733]: I1204 19:37:03.043551 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-a076-account-create-update-lsnb7"] Dec 04 19:37:04 crc kubenswrapper[4733]: I1204 19:37:04.357131 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8691a4ff-ced3-4098-a64d-c6d169d93eba" path="/var/lib/kubelet/pods/8691a4ff-ced3-4098-a64d-c6d169d93eba/volumes" Dec 04 19:37:15 crc kubenswrapper[4733]: I1204 19:37:15.362819 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:37:15 crc kubenswrapper[4733]: I1204 19:37:15.363494 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:37:16 crc kubenswrapper[4733]: I1204 19:37:16.055374 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-7hhh8"] Dec 04 19:37:16 crc kubenswrapper[4733]: I1204 19:37:16.063642 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-7hhh8"] Dec 04 19:37:16 crc kubenswrapper[4733]: I1204 19:37:16.362497 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58018da3-626d-4016-a9c9-4ff07a24ccf9" path="/var/lib/kubelet/pods/58018da3-626d-4016-a9c9-4ff07a24ccf9/volumes" Dec 04 19:37:27 crc kubenswrapper[4733]: I1204 19:37:27.675249 4733 scope.go:117] "RemoveContainer" containerID="7d37f601d15b0a66fc6301a369fbab32c1f52dfce3a3eb81b04d53cb4228476e" Dec 04 19:37:27 crc kubenswrapper[4733]: I1204 19:37:27.740660 4733 scope.go:117] "RemoveContainer" containerID="eed04ac06f54422891d78d3c542bc3ba195c6785e603bb08086b455165e8383f" Dec 04 19:37:27 crc kubenswrapper[4733]: I1204 19:37:27.776294 4733 scope.go:117] "RemoveContainer" containerID="4261675e3a05ddfe8f7078a08c53837038356425668c371d7b41c87164ba9d2d" Dec 04 19:37:27 crc kubenswrapper[4733]: I1204 19:37:27.813947 4733 scope.go:117] "RemoveContainer" containerID="ff3c098b1d71580667746addf9641c19911a6828a70aaa5093b72ad9c1a9ff52" Dec 04 19:37:27 crc kubenswrapper[4733]: I1204 19:37:27.851915 4733 scope.go:117] "RemoveContainer" containerID="97c793a24dcb898273e69f204c5adeac185a7c5755d3546329357a53a26c8b20" Dec 04 19:37:27 crc kubenswrapper[4733]: I1204 19:37:27.907009 4733 scope.go:117] "RemoveContainer" containerID="0822242418ead5fc50671a3368e47e07d7edd055a303357a076b0f92cb517633" Dec 04 19:37:45 crc kubenswrapper[4733]: I1204 19:37:45.361578 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:37:45 crc kubenswrapper[4733]: I1204 19:37:45.363166 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:38:15 crc kubenswrapper[4733]: I1204 19:38:15.362381 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:38:15 crc kubenswrapper[4733]: I1204 19:38:15.363050 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:38:15 crc kubenswrapper[4733]: I1204 19:38:15.363099 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:38:15 crc kubenswrapper[4733]: I1204 19:38:15.364204 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:38:15 crc kubenswrapper[4733]: I1204 19:38:15.364269 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" gracePeriod=600 Dec 04 19:38:15 crc kubenswrapper[4733]: E1204 19:38:15.514212 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:38:16 crc kubenswrapper[4733]: I1204 19:38:16.072663 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" exitCode=0 Dec 04 19:38:16 crc kubenswrapper[4733]: I1204 19:38:16.072770 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96"} Dec 04 19:38:16 crc kubenswrapper[4733]: I1204 19:38:16.073182 4733 scope.go:117] "RemoveContainer" containerID="1059a4cf6d36f458e18b39d02074ccfc6d3afb83e3b089fc9d8f384eed628898" Dec 04 19:38:16 crc kubenswrapper[4733]: I1204 19:38:16.074024 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:38:16 crc kubenswrapper[4733]: E1204 19:38:16.074470 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:38:31 crc kubenswrapper[4733]: I1204 19:38:31.336657 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:38:31 crc kubenswrapper[4733]: E1204 19:38:31.338208 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:38:46 crc kubenswrapper[4733]: I1204 19:38:46.336723 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:38:46 crc kubenswrapper[4733]: E1204 19:38:46.337515 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:38:57 crc kubenswrapper[4733]: I1204 19:38:57.336400 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:38:57 crc kubenswrapper[4733]: E1204 19:38:57.337256 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:39:08 crc kubenswrapper[4733]: I1204 19:39:08.341194 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:39:08 crc kubenswrapper[4733]: E1204 19:39:08.343439 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:39:20 crc kubenswrapper[4733]: I1204 19:39:20.337509 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:39:20 crc kubenswrapper[4733]: E1204 19:39:20.338781 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:39:33 crc kubenswrapper[4733]: I1204 19:39:33.337479 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:39:33 crc kubenswrapper[4733]: E1204 19:39:33.338730 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:39:46 crc kubenswrapper[4733]: I1204 19:39:46.337068 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:39:46 crc kubenswrapper[4733]: E1204 19:39:46.338488 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:39:58 crc kubenswrapper[4733]: I1204 19:39:58.345862 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:39:58 crc kubenswrapper[4733]: E1204 19:39:58.346737 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:40:13 crc kubenswrapper[4733]: I1204 19:40:13.335766 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:40:13 crc kubenswrapper[4733]: E1204 19:40:13.336893 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:40:13 crc kubenswrapper[4733]: I1204 19:40:13.388873 4733 generic.go:334] "Generic (PLEG): container finished" podID="3dacba4e-76b0-40f9-92b8-67feb437f1a7" containerID="c88c38519f002e57eb91ee0451a676a859188649e782903da78a89e088707da7" exitCode=0 Dec 04 19:40:13 crc kubenswrapper[4733]: I1204 19:40:13.388927 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" event={"ID":"3dacba4e-76b0-40f9-92b8-67feb437f1a7","Type":"ContainerDied","Data":"c88c38519f002e57eb91ee0451a676a859188649e782903da78a89e088707da7"} Dec 04 19:40:14 crc kubenswrapper[4733]: I1204 19:40:14.839990 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.004135 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhrsw\" (UniqueName: \"kubernetes.io/projected/3dacba4e-76b0-40f9-92b8-67feb437f1a7-kube-api-access-bhrsw\") pod \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.004248 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-inventory\") pod \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.004302 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ssh-key\") pod \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.004382 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ceph\") pod \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.004492 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-tripleo-cleanup-combined-ca-bundle\") pod \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\" (UID: \"3dacba4e-76b0-40f9-92b8-67feb437f1a7\") " Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.009915 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ceph" (OuterVolumeSpecName: "ceph") pod "3dacba4e-76b0-40f9-92b8-67feb437f1a7" (UID: "3dacba4e-76b0-40f9-92b8-67feb437f1a7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.010159 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "3dacba4e-76b0-40f9-92b8-67feb437f1a7" (UID: "3dacba4e-76b0-40f9-92b8-67feb437f1a7"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.018508 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dacba4e-76b0-40f9-92b8-67feb437f1a7-kube-api-access-bhrsw" (OuterVolumeSpecName: "kube-api-access-bhrsw") pod "3dacba4e-76b0-40f9-92b8-67feb437f1a7" (UID: "3dacba4e-76b0-40f9-92b8-67feb437f1a7"). InnerVolumeSpecName "kube-api-access-bhrsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.043757 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3dacba4e-76b0-40f9-92b8-67feb437f1a7" (UID: "3dacba4e-76b0-40f9-92b8-67feb437f1a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.053791 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-inventory" (OuterVolumeSpecName: "inventory") pod "3dacba4e-76b0-40f9-92b8-67feb437f1a7" (UID: "3dacba4e-76b0-40f9-92b8-67feb437f1a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.106804 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.106840 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.106851 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.106864 4733 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dacba4e-76b0-40f9-92b8-67feb437f1a7-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.106877 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhrsw\" (UniqueName: \"kubernetes.io/projected/3dacba4e-76b0-40f9-92b8-67feb437f1a7-kube-api-access-bhrsw\") on node \"crc\" DevicePath \"\"" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.409914 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" event={"ID":"3dacba4e-76b0-40f9-92b8-67feb437f1a7","Type":"ContainerDied","Data":"728c39afcb66cc16ed43a83df26af29cab638cba9efcaec754a2951883c2b1ac"} Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.409986 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="728c39afcb66cc16ed43a83df26af29cab638cba9efcaec754a2951883c2b1ac" Dec 04 19:40:15 crc kubenswrapper[4733]: I1204 19:40:15.410050 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.045236 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-7q5rg"] Dec 04 19:40:18 crc kubenswrapper[4733]: E1204 19:40:18.046746 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerName="registry-server" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.046765 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerName="registry-server" Dec 04 19:40:18 crc kubenswrapper[4733]: E1204 19:40:18.046785 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dacba4e-76b0-40f9-92b8-67feb437f1a7" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.046821 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dacba4e-76b0-40f9-92b8-67feb437f1a7" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 04 19:40:18 crc kubenswrapper[4733]: E1204 19:40:18.046847 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerName="extract-content" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.046854 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerName="extract-content" Dec 04 19:40:18 crc kubenswrapper[4733]: E1204 19:40:18.046866 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerName="extract-utilities" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.046873 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerName="extract-utilities" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.047240 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dd98991-e859-4aa3-a3b0-dc2ed5ce31c7" containerName="registry-server" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.047272 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dacba4e-76b0-40f9-92b8-67feb437f1a7" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.048552 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.050874 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.052908 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.057782 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.058200 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.058437 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-7q5rg"] Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.180017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-inventory\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.180445 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq656\" (UniqueName: \"kubernetes.io/projected/f85c314e-3306-4fa3-a717-9f31c1d4817f-kube-api-access-sq656\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.180564 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.180735 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ceph\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.180973 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.283275 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.283380 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-inventory\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.283593 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq656\" (UniqueName: \"kubernetes.io/projected/f85c314e-3306-4fa3-a717-9f31c1d4817f-kube-api-access-sq656\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.283657 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.283789 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ceph\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.291268 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.291285 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.293745 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ceph\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.301939 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-inventory\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.305329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq656\" (UniqueName: \"kubernetes.io/projected/f85c314e-3306-4fa3-a717-9f31c1d4817f-kube-api-access-sq656\") pod \"bootstrap-openstack-openstack-cell1-7q5rg\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:18 crc kubenswrapper[4733]: I1204 19:40:18.370178 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:40:19 crc kubenswrapper[4733]: I1204 19:40:19.045208 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-7q5rg"] Dec 04 19:40:19 crc kubenswrapper[4733]: I1204 19:40:19.452301 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" event={"ID":"f85c314e-3306-4fa3-a717-9f31c1d4817f","Type":"ContainerStarted","Data":"3e18a83fc6de9d2dfaf29d7741ef6ff255195679a6142e99da18ea53eaab7ecf"} Dec 04 19:40:20 crc kubenswrapper[4733]: I1204 19:40:20.465714 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" event={"ID":"f85c314e-3306-4fa3-a717-9f31c1d4817f","Type":"ContainerStarted","Data":"ded808edfaf10077047448fdd37e2a9aebdcdfac9537c2456e0c758366a12edc"} Dec 04 19:40:20 crc kubenswrapper[4733]: I1204 19:40:20.495052 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" podStartSLOduration=2.012324696 podStartE2EDuration="2.495028693s" podCreationTimestamp="2025-12-04 19:40:18 +0000 UTC" firstStartedPulling="2025-12-04 19:40:19.048387482 +0000 UTC m=+7281.003748528" lastFinishedPulling="2025-12-04 19:40:19.531091479 +0000 UTC m=+7281.486452525" observedRunningTime="2025-12-04 19:40:20.485132235 +0000 UTC m=+7282.440493281" watchObservedRunningTime="2025-12-04 19:40:20.495028693 +0000 UTC m=+7282.450389739" Dec 04 19:40:28 crc kubenswrapper[4733]: I1204 19:40:28.348423 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:40:28 crc kubenswrapper[4733]: E1204 19:40:28.350106 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:40:39 crc kubenswrapper[4733]: I1204 19:40:39.336297 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:40:39 crc kubenswrapper[4733]: E1204 19:40:39.337370 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:40:51 crc kubenswrapper[4733]: I1204 19:40:51.336159 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:40:51 crc kubenswrapper[4733]: E1204 19:40:51.337043 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:41:05 crc kubenswrapper[4733]: I1204 19:41:05.336385 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:41:05 crc kubenswrapper[4733]: E1204 19:41:05.337104 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:41:09 crc kubenswrapper[4733]: I1204 19:41:09.877549 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7vmjq"] Dec 04 19:41:09 crc kubenswrapper[4733]: I1204 19:41:09.881153 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:09 crc kubenswrapper[4733]: I1204 19:41:09.894029 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7vmjq"] Dec 04 19:41:09 crc kubenswrapper[4733]: I1204 19:41:09.942848 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-catalog-content\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:09 crc kubenswrapper[4733]: I1204 19:41:09.943242 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-utilities\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:09 crc kubenswrapper[4733]: I1204 19:41:09.943607 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvmtc\" (UniqueName: \"kubernetes.io/projected/d7ac8d66-023a-4989-a3ad-62482beb382e-kube-api-access-jvmtc\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:10 crc kubenswrapper[4733]: I1204 19:41:10.045618 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvmtc\" (UniqueName: \"kubernetes.io/projected/d7ac8d66-023a-4989-a3ad-62482beb382e-kube-api-access-jvmtc\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:10 crc kubenswrapper[4733]: I1204 19:41:10.045697 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-catalog-content\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:10 crc kubenswrapper[4733]: I1204 19:41:10.045734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-utilities\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:10 crc kubenswrapper[4733]: I1204 19:41:10.046356 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-catalog-content\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:10 crc kubenswrapper[4733]: I1204 19:41:10.046384 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-utilities\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:10 crc kubenswrapper[4733]: I1204 19:41:10.066629 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvmtc\" (UniqueName: \"kubernetes.io/projected/d7ac8d66-023a-4989-a3ad-62482beb382e-kube-api-access-jvmtc\") pod \"certified-operators-7vmjq\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:10 crc kubenswrapper[4733]: I1204 19:41:10.212961 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:10 crc kubenswrapper[4733]: I1204 19:41:10.694291 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7vmjq"] Dec 04 19:41:11 crc kubenswrapper[4733]: I1204 19:41:11.025463 4733 generic.go:334] "Generic (PLEG): container finished" podID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerID="fda5997a954620b5f0eef7bd8da730562e3f25a363a2d5ae1ee25e4f75715242" exitCode=0 Dec 04 19:41:11 crc kubenswrapper[4733]: I1204 19:41:11.025503 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7vmjq" event={"ID":"d7ac8d66-023a-4989-a3ad-62482beb382e","Type":"ContainerDied","Data":"fda5997a954620b5f0eef7bd8da730562e3f25a363a2d5ae1ee25e4f75715242"} Dec 04 19:41:11 crc kubenswrapper[4733]: I1204 19:41:11.025710 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7vmjq" event={"ID":"d7ac8d66-023a-4989-a3ad-62482beb382e","Type":"ContainerStarted","Data":"08edb36fd8a65d41121790976d1c7bbd124111e021bfab0b3e8a5b94eb190092"} Dec 04 19:41:11 crc kubenswrapper[4733]: I1204 19:41:11.027393 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:41:12 crc kubenswrapper[4733]: I1204 19:41:12.038542 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7vmjq" event={"ID":"d7ac8d66-023a-4989-a3ad-62482beb382e","Type":"ContainerStarted","Data":"26439a8495ca2993ce216c8c8abf6136904fc14aac585b1dfc5b4c09521fa54e"} Dec 04 19:41:13 crc kubenswrapper[4733]: I1204 19:41:13.061245 4733 generic.go:334] "Generic (PLEG): container finished" podID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerID="26439a8495ca2993ce216c8c8abf6136904fc14aac585b1dfc5b4c09521fa54e" exitCode=0 Dec 04 19:41:13 crc kubenswrapper[4733]: I1204 19:41:13.061462 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7vmjq" event={"ID":"d7ac8d66-023a-4989-a3ad-62482beb382e","Type":"ContainerDied","Data":"26439a8495ca2993ce216c8c8abf6136904fc14aac585b1dfc5b4c09521fa54e"} Dec 04 19:41:14 crc kubenswrapper[4733]: I1204 19:41:14.081033 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7vmjq" event={"ID":"d7ac8d66-023a-4989-a3ad-62482beb382e","Type":"ContainerStarted","Data":"72aaddc5b841f7123b30a7d619522d22916424c1f803ac29e6e82d907d94cece"} Dec 04 19:41:14 crc kubenswrapper[4733]: I1204 19:41:14.108615 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7vmjq" podStartSLOduration=2.651094459 podStartE2EDuration="5.108595035s" podCreationTimestamp="2025-12-04 19:41:09 +0000 UTC" firstStartedPulling="2025-12-04 19:41:11.027125375 +0000 UTC m=+7332.982486421" lastFinishedPulling="2025-12-04 19:41:13.484625941 +0000 UTC m=+7335.439986997" observedRunningTime="2025-12-04 19:41:14.103921758 +0000 UTC m=+7336.059282844" watchObservedRunningTime="2025-12-04 19:41:14.108595035 +0000 UTC m=+7336.063956091" Dec 04 19:41:20 crc kubenswrapper[4733]: I1204 19:41:20.213209 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:20 crc kubenswrapper[4733]: I1204 19:41:20.213979 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:20 crc kubenswrapper[4733]: I1204 19:41:20.284238 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:20 crc kubenswrapper[4733]: I1204 19:41:20.335977 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:41:20 crc kubenswrapper[4733]: E1204 19:41:20.336539 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:41:21 crc kubenswrapper[4733]: I1204 19:41:21.244093 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:21 crc kubenswrapper[4733]: I1204 19:41:21.309252 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7vmjq"] Dec 04 19:41:23 crc kubenswrapper[4733]: I1204 19:41:23.180057 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7vmjq" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerName="registry-server" containerID="cri-o://72aaddc5b841f7123b30a7d619522d22916424c1f803ac29e6e82d907d94cece" gracePeriod=2 Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.225320 4733 generic.go:334] "Generic (PLEG): container finished" podID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerID="72aaddc5b841f7123b30a7d619522d22916424c1f803ac29e6e82d907d94cece" exitCode=0 Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.225410 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7vmjq" event={"ID":"d7ac8d66-023a-4989-a3ad-62482beb382e","Type":"ContainerDied","Data":"72aaddc5b841f7123b30a7d619522d22916424c1f803ac29e6e82d907d94cece"} Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.225998 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7vmjq" event={"ID":"d7ac8d66-023a-4989-a3ad-62482beb382e","Type":"ContainerDied","Data":"08edb36fd8a65d41121790976d1c7bbd124111e021bfab0b3e8a5b94eb190092"} Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.226024 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08edb36fd8a65d41121790976d1c7bbd124111e021bfab0b3e8a5b94eb190092" Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.297286 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.389508 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-utilities\") pod \"d7ac8d66-023a-4989-a3ad-62482beb382e\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.389628 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-catalog-content\") pod \"d7ac8d66-023a-4989-a3ad-62482beb382e\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.389900 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvmtc\" (UniqueName: \"kubernetes.io/projected/d7ac8d66-023a-4989-a3ad-62482beb382e-kube-api-access-jvmtc\") pod \"d7ac8d66-023a-4989-a3ad-62482beb382e\" (UID: \"d7ac8d66-023a-4989-a3ad-62482beb382e\") " Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.390879 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-utilities" (OuterVolumeSpecName: "utilities") pod "d7ac8d66-023a-4989-a3ad-62482beb382e" (UID: "d7ac8d66-023a-4989-a3ad-62482beb382e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.398997 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7ac8d66-023a-4989-a3ad-62482beb382e-kube-api-access-jvmtc" (OuterVolumeSpecName: "kube-api-access-jvmtc") pod "d7ac8d66-023a-4989-a3ad-62482beb382e" (UID: "d7ac8d66-023a-4989-a3ad-62482beb382e"). InnerVolumeSpecName "kube-api-access-jvmtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.438019 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7ac8d66-023a-4989-a3ad-62482beb382e" (UID: "d7ac8d66-023a-4989-a3ad-62482beb382e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.492483 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.492683 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvmtc\" (UniqueName: \"kubernetes.io/projected/d7ac8d66-023a-4989-a3ad-62482beb382e-kube-api-access-jvmtc\") on node \"crc\" DevicePath \"\"" Dec 04 19:41:24 crc kubenswrapper[4733]: I1204 19:41:24.492719 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ac8d66-023a-4989-a3ad-62482beb382e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:41:25 crc kubenswrapper[4733]: I1204 19:41:25.239717 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7vmjq" Dec 04 19:41:25 crc kubenswrapper[4733]: I1204 19:41:25.308565 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7vmjq"] Dec 04 19:41:25 crc kubenswrapper[4733]: I1204 19:41:25.317543 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7vmjq"] Dec 04 19:41:26 crc kubenswrapper[4733]: I1204 19:41:26.347441 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" path="/var/lib/kubelet/pods/d7ac8d66-023a-4989-a3ad-62482beb382e/volumes" Dec 04 19:41:33 crc kubenswrapper[4733]: I1204 19:41:33.338293 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:41:33 crc kubenswrapper[4733]: E1204 19:41:33.340952 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:41:45 crc kubenswrapper[4733]: I1204 19:41:45.336635 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:41:45 crc kubenswrapper[4733]: E1204 19:41:45.337577 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:41:59 crc kubenswrapper[4733]: I1204 19:41:59.335453 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:41:59 crc kubenswrapper[4733]: E1204 19:41:59.336367 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:42:12 crc kubenswrapper[4733]: I1204 19:42:12.336289 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:42:12 crc kubenswrapper[4733]: E1204 19:42:12.337234 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:42:26 crc kubenswrapper[4733]: I1204 19:42:26.337914 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:42:26 crc kubenswrapper[4733]: E1204 19:42:26.339070 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:42:39 crc kubenswrapper[4733]: I1204 19:42:39.336267 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:42:39 crc kubenswrapper[4733]: E1204 19:42:39.337094 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:42:53 crc kubenswrapper[4733]: I1204 19:42:53.335982 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:42:53 crc kubenswrapper[4733]: E1204 19:42:53.336811 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:43:08 crc kubenswrapper[4733]: I1204 19:43:08.341491 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:43:08 crc kubenswrapper[4733]: E1204 19:43:08.342260 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:43:21 crc kubenswrapper[4733]: I1204 19:43:21.336754 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:43:22 crc kubenswrapper[4733]: I1204 19:43:22.561460 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"6702c9ff250314030e3d19e8ac891392736fe3f613350f0fa8b792973e875133"} Dec 04 19:43:31 crc kubenswrapper[4733]: I1204 19:43:31.656978 4733 generic.go:334] "Generic (PLEG): container finished" podID="f85c314e-3306-4fa3-a717-9f31c1d4817f" containerID="ded808edfaf10077047448fdd37e2a9aebdcdfac9537c2456e0c758366a12edc" exitCode=0 Dec 04 19:43:31 crc kubenswrapper[4733]: I1204 19:43:31.657062 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" event={"ID":"f85c314e-3306-4fa3-a717-9f31c1d4817f","Type":"ContainerDied","Data":"ded808edfaf10077047448fdd37e2a9aebdcdfac9537c2456e0c758366a12edc"} Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.119573 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.212209 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq656\" (UniqueName: \"kubernetes.io/projected/f85c314e-3306-4fa3-a717-9f31c1d4817f-kube-api-access-sq656\") pod \"f85c314e-3306-4fa3-a717-9f31c1d4817f\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.212548 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-bootstrap-combined-ca-bundle\") pod \"f85c314e-3306-4fa3-a717-9f31c1d4817f\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.212761 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-inventory\") pod \"f85c314e-3306-4fa3-a717-9f31c1d4817f\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.212923 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ceph\") pod \"f85c314e-3306-4fa3-a717-9f31c1d4817f\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.213107 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ssh-key\") pod \"f85c314e-3306-4fa3-a717-9f31c1d4817f\" (UID: \"f85c314e-3306-4fa3-a717-9f31c1d4817f\") " Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.219491 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f85c314e-3306-4fa3-a717-9f31c1d4817f" (UID: "f85c314e-3306-4fa3-a717-9f31c1d4817f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.219518 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ceph" (OuterVolumeSpecName: "ceph") pod "f85c314e-3306-4fa3-a717-9f31c1d4817f" (UID: "f85c314e-3306-4fa3-a717-9f31c1d4817f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.219552 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f85c314e-3306-4fa3-a717-9f31c1d4817f-kube-api-access-sq656" (OuterVolumeSpecName: "kube-api-access-sq656") pod "f85c314e-3306-4fa3-a717-9f31c1d4817f" (UID: "f85c314e-3306-4fa3-a717-9f31c1d4817f"). InnerVolumeSpecName "kube-api-access-sq656". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.243101 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-inventory" (OuterVolumeSpecName: "inventory") pod "f85c314e-3306-4fa3-a717-9f31c1d4817f" (UID: "f85c314e-3306-4fa3-a717-9f31c1d4817f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.248575 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f85c314e-3306-4fa3-a717-9f31c1d4817f" (UID: "f85c314e-3306-4fa3-a717-9f31c1d4817f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.316195 4733 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.316229 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.316241 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.316249 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f85c314e-3306-4fa3-a717-9f31c1d4817f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.316258 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq656\" (UniqueName: \"kubernetes.io/projected/f85c314e-3306-4fa3-a717-9f31c1d4817f-kube-api-access-sq656\") on node \"crc\" DevicePath \"\"" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.681345 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" event={"ID":"f85c314e-3306-4fa3-a717-9f31c1d4817f","Type":"ContainerDied","Data":"3e18a83fc6de9d2dfaf29d7741ef6ff255195679a6142e99da18ea53eaab7ecf"} Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.681394 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e18a83fc6de9d2dfaf29d7741ef6ff255195679a6142e99da18ea53eaab7ecf" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.681455 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-7q5rg" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.763637 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-s9lm6"] Dec 04 19:43:33 crc kubenswrapper[4733]: E1204 19:43:33.764182 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerName="extract-content" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.764211 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerName="extract-content" Dec 04 19:43:33 crc kubenswrapper[4733]: E1204 19:43:33.764300 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerName="extract-utilities" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.764311 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerName="extract-utilities" Dec 04 19:43:33 crc kubenswrapper[4733]: E1204 19:43:33.764333 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerName="registry-server" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.764341 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerName="registry-server" Dec 04 19:43:33 crc kubenswrapper[4733]: E1204 19:43:33.764369 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85c314e-3306-4fa3-a717-9f31c1d4817f" containerName="bootstrap-openstack-openstack-cell1" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.764377 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85c314e-3306-4fa3-a717-9f31c1d4817f" containerName="bootstrap-openstack-openstack-cell1" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.764645 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7ac8d66-023a-4989-a3ad-62482beb382e" containerName="registry-server" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.764679 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85c314e-3306-4fa3-a717-9f31c1d4817f" containerName="bootstrap-openstack-openstack-cell1" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.765560 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.770831 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.771035 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.771059 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.774670 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.780530 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-s9lm6"] Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.931494 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qm64\" (UniqueName: \"kubernetes.io/projected/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-kube-api-access-5qm64\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.931756 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ssh-key\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.932121 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-inventory\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:33 crc kubenswrapper[4733]: I1204 19:43:33.932305 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ceph\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.034950 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qm64\" (UniqueName: \"kubernetes.io/projected/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-kube-api-access-5qm64\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.035714 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ssh-key\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.035998 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-inventory\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.036127 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ceph\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.041770 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-inventory\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.056724 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ceph\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.058588 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ssh-key\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.059200 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qm64\" (UniqueName: \"kubernetes.io/projected/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-kube-api-access-5qm64\") pod \"download-cache-openstack-openstack-cell1-s9lm6\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.087323 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.623438 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-s9lm6"] Dec 04 19:43:34 crc kubenswrapper[4733]: I1204 19:43:34.693120 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" event={"ID":"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79","Type":"ContainerStarted","Data":"836a0ae1dbea262eff566a59898f7f642848d2548d465820da7cbf9f5dd4a0a9"} Dec 04 19:43:35 crc kubenswrapper[4733]: I1204 19:43:35.704141 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" event={"ID":"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79","Type":"ContainerStarted","Data":"284fec4c4e5df306d393c8afe095153a0cf66ce47126b3b59e53f5b1b4286f1c"} Dec 04 19:43:35 crc kubenswrapper[4733]: I1204 19:43:35.731051 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" podStartSLOduration=2.28399324 podStartE2EDuration="2.73103659s" podCreationTimestamp="2025-12-04 19:43:33 +0000 UTC" firstStartedPulling="2025-12-04 19:43:34.633688663 +0000 UTC m=+7476.589049729" lastFinishedPulling="2025-12-04 19:43:35.080732033 +0000 UTC m=+7477.036093079" observedRunningTime="2025-12-04 19:43:35.730732642 +0000 UTC m=+7477.686093698" watchObservedRunningTime="2025-12-04 19:43:35.73103659 +0000 UTC m=+7477.686397636" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.183615 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn"] Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.185684 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.187382 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.190188 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.212245 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn"] Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.305614 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzt7b\" (UniqueName: \"kubernetes.io/projected/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-kube-api-access-lzt7b\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.305674 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-secret-volume\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.305736 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-config-volume\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.408295 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzt7b\" (UniqueName: \"kubernetes.io/projected/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-kube-api-access-lzt7b\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.408364 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-secret-volume\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.408393 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-config-volume\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.409554 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-config-volume\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.420735 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-secret-volume\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.435631 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzt7b\" (UniqueName: \"kubernetes.io/projected/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-kube-api-access-lzt7b\") pod \"collect-profiles-29414625-wg9xn\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:00 crc kubenswrapper[4733]: I1204 19:45:00.520171 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:01 crc kubenswrapper[4733]: I1204 19:45:01.226990 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn"] Dec 04 19:45:01 crc kubenswrapper[4733]: I1204 19:45:01.653338 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" event={"ID":"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e","Type":"ContainerStarted","Data":"a1d0df7b29d4491e5bbacd2591672d81084bfd407579d9aa4fb00e8515a7d427"} Dec 04 19:45:01 crc kubenswrapper[4733]: I1204 19:45:01.653675 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" event={"ID":"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e","Type":"ContainerStarted","Data":"cf5096b2dd131d8aad31e64bb103c5c4f2149c56152e5c47a9ce37ebd7064764"} Dec 04 19:45:01 crc kubenswrapper[4733]: I1204 19:45:01.676364 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" podStartSLOduration=1.6763481439999999 podStartE2EDuration="1.676348144s" podCreationTimestamp="2025-12-04 19:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 19:45:01.671510663 +0000 UTC m=+7563.626871729" watchObservedRunningTime="2025-12-04 19:45:01.676348144 +0000 UTC m=+7563.631709190" Dec 04 19:45:02 crc kubenswrapper[4733]: I1204 19:45:02.662849 4733 generic.go:334] "Generic (PLEG): container finished" podID="14e5bf5e-eafc-4c9a-9a41-e1608463cf0e" containerID="a1d0df7b29d4491e5bbacd2591672d81084bfd407579d9aa4fb00e8515a7d427" exitCode=0 Dec 04 19:45:02 crc kubenswrapper[4733]: I1204 19:45:02.662904 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" event={"ID":"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e","Type":"ContainerDied","Data":"a1d0df7b29d4491e5bbacd2591672d81084bfd407579d9aa4fb00e8515a7d427"} Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.214934 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.310093 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7"] Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.320725 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414580-br8z7"] Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.350253 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2beee50f-af17-48d4-9224-b34248a31cfd" path="/var/lib/kubelet/pods/2beee50f-af17-48d4-9224-b34248a31cfd/volumes" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.397646 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-config-volume\") pod \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.397870 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-secret-volume\") pod \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.398011 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzt7b\" (UniqueName: \"kubernetes.io/projected/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-kube-api-access-lzt7b\") pod \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\" (UID: \"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e\") " Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.398635 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-config-volume" (OuterVolumeSpecName: "config-volume") pod "14e5bf5e-eafc-4c9a-9a41-e1608463cf0e" (UID: "14e5bf5e-eafc-4c9a-9a41-e1608463cf0e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.399038 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.403882 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "14e5bf5e-eafc-4c9a-9a41-e1608463cf0e" (UID: "14e5bf5e-eafc-4c9a-9a41-e1608463cf0e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.405014 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-kube-api-access-lzt7b" (OuterVolumeSpecName: "kube-api-access-lzt7b") pod "14e5bf5e-eafc-4c9a-9a41-e1608463cf0e" (UID: "14e5bf5e-eafc-4c9a-9a41-e1608463cf0e"). InnerVolumeSpecName "kube-api-access-lzt7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.501074 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.501102 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzt7b\" (UniqueName: \"kubernetes.io/projected/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e-kube-api-access-lzt7b\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.693446 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" event={"ID":"14e5bf5e-eafc-4c9a-9a41-e1608463cf0e","Type":"ContainerDied","Data":"cf5096b2dd131d8aad31e64bb103c5c4f2149c56152e5c47a9ce37ebd7064764"} Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.693766 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf5096b2dd131d8aad31e64bb103c5c4f2149c56152e5c47a9ce37ebd7064764" Dec 04 19:45:04 crc kubenswrapper[4733]: I1204 19:45:04.693577 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn" Dec 04 19:45:09 crc kubenswrapper[4733]: I1204 19:45:09.743892 4733 generic.go:334] "Generic (PLEG): container finished" podID="b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79" containerID="284fec4c4e5df306d393c8afe095153a0cf66ce47126b3b59e53f5b1b4286f1c" exitCode=0 Dec 04 19:45:09 crc kubenswrapper[4733]: I1204 19:45:09.744542 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" event={"ID":"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79","Type":"ContainerDied","Data":"284fec4c4e5df306d393c8afe095153a0cf66ce47126b3b59e53f5b1b4286f1c"} Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.409527 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.582759 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ssh-key\") pod \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.582845 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ceph\") pod \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.582951 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-inventory\") pod \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.583012 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qm64\" (UniqueName: \"kubernetes.io/projected/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-kube-api-access-5qm64\") pod \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\" (UID: \"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79\") " Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.589374 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-kube-api-access-5qm64" (OuterVolumeSpecName: "kube-api-access-5qm64") pod "b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79" (UID: "b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79"). InnerVolumeSpecName "kube-api-access-5qm64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.594030 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ceph" (OuterVolumeSpecName: "ceph") pod "b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79" (UID: "b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.615549 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-inventory" (OuterVolumeSpecName: "inventory") pod "b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79" (UID: "b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.616228 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79" (UID: "b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.685785 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.685844 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.685854 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.685863 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qm64\" (UniqueName: \"kubernetes.io/projected/b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79-kube-api-access-5qm64\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.767010 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" event={"ID":"b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79","Type":"ContainerDied","Data":"836a0ae1dbea262eff566a59898f7f642848d2548d465820da7cbf9f5dd4a0a9"} Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.767255 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="836a0ae1dbea262eff566a59898f7f642848d2548d465820da7cbf9f5dd4a0a9" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.767062 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-s9lm6" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.840352 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-9v588"] Dec 04 19:45:11 crc kubenswrapper[4733]: E1204 19:45:11.845943 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79" containerName="download-cache-openstack-openstack-cell1" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.845972 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79" containerName="download-cache-openstack-openstack-cell1" Dec 04 19:45:11 crc kubenswrapper[4733]: E1204 19:45:11.846003 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14e5bf5e-eafc-4c9a-9a41-e1608463cf0e" containerName="collect-profiles" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.846010 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="14e5bf5e-eafc-4c9a-9a41-e1608463cf0e" containerName="collect-profiles" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.846196 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="14e5bf5e-eafc-4c9a-9a41-e1608463cf0e" containerName="collect-profiles" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.846222 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79" containerName="download-cache-openstack-openstack-cell1" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.847006 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.848708 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.849433 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.849569 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.849676 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.849914 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-9v588"] Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.992146 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ssh-key\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.992260 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-inventory\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.992299 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prh9h\" (UniqueName: \"kubernetes.io/projected/1b7b4789-500e-40d1-bddd-332ab22536e9-kube-api-access-prh9h\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:11 crc kubenswrapper[4733]: I1204 19:45:11.992406 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ceph\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.094817 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ssh-key\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.095107 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-inventory\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.095198 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prh9h\" (UniqueName: \"kubernetes.io/projected/1b7b4789-500e-40d1-bddd-332ab22536e9-kube-api-access-prh9h\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.095364 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ceph\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.100509 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ceph\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.100679 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ssh-key\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.101249 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-inventory\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.116917 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prh9h\" (UniqueName: \"kubernetes.io/projected/1b7b4789-500e-40d1-bddd-332ab22536e9-kube-api-access-prh9h\") pod \"configure-network-openstack-openstack-cell1-9v588\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.168819 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.744266 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-9v588"] Dec 04 19:45:12 crc kubenswrapper[4733]: I1204 19:45:12.775614 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-9v588" event={"ID":"1b7b4789-500e-40d1-bddd-332ab22536e9","Type":"ContainerStarted","Data":"74544c3e3b76ed7240b38b73eb37a206687cbf4b0ba4cb76d597b7b91075b160"} Dec 04 19:45:13 crc kubenswrapper[4733]: I1204 19:45:13.787443 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-9v588" event={"ID":"1b7b4789-500e-40d1-bddd-332ab22536e9","Type":"ContainerStarted","Data":"613a370d675ce7c1d82cb5b42d705ff072705cf2c8586f051107e1d86bb746c3"} Dec 04 19:45:13 crc kubenswrapper[4733]: I1204 19:45:13.808195 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-9v588" podStartSLOduration=2.364975641 podStartE2EDuration="2.808176288s" podCreationTimestamp="2025-12-04 19:45:11 +0000 UTC" firstStartedPulling="2025-12-04 19:45:12.753037793 +0000 UTC m=+7574.708398839" lastFinishedPulling="2025-12-04 19:45:13.1962384 +0000 UTC m=+7575.151599486" observedRunningTime="2025-12-04 19:45:13.802262328 +0000 UTC m=+7575.757623394" watchObservedRunningTime="2025-12-04 19:45:13.808176288 +0000 UTC m=+7575.763537334" Dec 04 19:45:28 crc kubenswrapper[4733]: I1204 19:45:28.241097 4733 scope.go:117] "RemoveContainer" containerID="98399f88ed3c435260e57a8e3a6c633977de3439f9edb5927487c36eab94b7c7" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.378002 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ck2l5"] Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.382320 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.392182 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ck2l5"] Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.494040 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8j8v\" (UniqueName: \"kubernetes.io/projected/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-kube-api-access-l8j8v\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.494519 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-utilities\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.494823 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-catalog-content\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.596263 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-catalog-content\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.596370 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8j8v\" (UniqueName: \"kubernetes.io/projected/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-kube-api-access-l8j8v\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.596438 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-utilities\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.596938 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-utilities\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.597122 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-catalog-content\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.629088 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8j8v\" (UniqueName: \"kubernetes.io/projected/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-kube-api-access-l8j8v\") pod \"community-operators-ck2l5\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:34 crc kubenswrapper[4733]: I1204 19:45:34.705387 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:35 crc kubenswrapper[4733]: I1204 19:45:35.201456 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ck2l5"] Dec 04 19:45:36 crc kubenswrapper[4733]: I1204 19:45:36.039502 4733 generic.go:334] "Generic (PLEG): container finished" podID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerID="4d030d57599e004aba4d8c10008c4ce7d34161e0a816576b21cc4b45715764e6" exitCode=0 Dec 04 19:45:36 crc kubenswrapper[4733]: I1204 19:45:36.039695 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck2l5" event={"ID":"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee","Type":"ContainerDied","Data":"4d030d57599e004aba4d8c10008c4ce7d34161e0a816576b21cc4b45715764e6"} Dec 04 19:45:36 crc kubenswrapper[4733]: I1204 19:45:36.039851 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck2l5" event={"ID":"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee","Type":"ContainerStarted","Data":"7741e9818892840694acef995d66d84577e992bfea810f1faa53be2e439435d8"} Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.053500 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck2l5" event={"ID":"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee","Type":"ContainerStarted","Data":"d4e91f2b86ea6a0ae45c88e9e59bf45a13d2718931097b55897034cf9a4bf140"} Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.772589 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-68b76"] Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.778717 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.785315 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68b76"] Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.865728 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-utilities\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.866056 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrgdb\" (UniqueName: \"kubernetes.io/projected/ada6064a-861f-448c-9142-a3c1e6fc0f02-kube-api-access-jrgdb\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.866258 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-catalog-content\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.969223 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-catalog-content\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.968430 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-catalog-content\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.969538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-utilities\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.969916 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-utilities\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.969570 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrgdb\" (UniqueName: \"kubernetes.io/projected/ada6064a-861f-448c-9142-a3c1e6fc0f02-kube-api-access-jrgdb\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:37 crc kubenswrapper[4733]: I1204 19:45:37.992222 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrgdb\" (UniqueName: \"kubernetes.io/projected/ada6064a-861f-448c-9142-a3c1e6fc0f02-kube-api-access-jrgdb\") pod \"redhat-marketplace-68b76\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:38 crc kubenswrapper[4733]: I1204 19:45:38.112043 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:38 crc kubenswrapper[4733]: W1204 19:45:38.817236 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podada6064a_861f_448c_9142_a3c1e6fc0f02.slice/crio-bde8dfc193067c585fc92e9effc1f758dd9d1efdbd8941f4857871797d590153 WatchSource:0}: Error finding container bde8dfc193067c585fc92e9effc1f758dd9d1efdbd8941f4857871797d590153: Status 404 returned error can't find the container with id bde8dfc193067c585fc92e9effc1f758dd9d1efdbd8941f4857871797d590153 Dec 04 19:45:38 crc kubenswrapper[4733]: I1204 19:45:38.818224 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68b76"] Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.078520 4733 generic.go:334] "Generic (PLEG): container finished" podID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerID="d4e91f2b86ea6a0ae45c88e9e59bf45a13d2718931097b55897034cf9a4bf140" exitCode=0 Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.078617 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck2l5" event={"ID":"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee","Type":"ContainerDied","Data":"d4e91f2b86ea6a0ae45c88e9e59bf45a13d2718931097b55897034cf9a4bf140"} Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.080836 4733 generic.go:334] "Generic (PLEG): container finished" podID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerID="e30f0ed8b26c67cd64fff15a972d0b369934e7fd3e6b3185bc4b01e9ca63ae92" exitCode=0 Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.080883 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68b76" event={"ID":"ada6064a-861f-448c-9142-a3c1e6fc0f02","Type":"ContainerDied","Data":"e30f0ed8b26c67cd64fff15a972d0b369934e7fd3e6b3185bc4b01e9ca63ae92"} Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.080917 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68b76" event={"ID":"ada6064a-861f-448c-9142-a3c1e6fc0f02","Type":"ContainerStarted","Data":"bde8dfc193067c585fc92e9effc1f758dd9d1efdbd8941f4857871797d590153"} Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.371169 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tvh6f"] Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.373550 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.393126 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tvh6f"] Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.410439 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-utilities\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.410503 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-catalog-content\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.410534 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttbhz\" (UniqueName: \"kubernetes.io/projected/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-kube-api-access-ttbhz\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.512729 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-utilities\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.512830 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-catalog-content\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.512865 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttbhz\" (UniqueName: \"kubernetes.io/projected/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-kube-api-access-ttbhz\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.513370 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-utilities\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.513407 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-catalog-content\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.545348 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttbhz\" (UniqueName: \"kubernetes.io/projected/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-kube-api-access-ttbhz\") pod \"redhat-operators-tvh6f\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:39 crc kubenswrapper[4733]: I1204 19:45:39.695249 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:40 crc kubenswrapper[4733]: I1204 19:45:40.092094 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68b76" event={"ID":"ada6064a-861f-448c-9142-a3c1e6fc0f02","Type":"ContainerStarted","Data":"97d73bd59b14b21a92499f91c1549c5920a3d108b907c38262665b977c111135"} Dec 04 19:45:40 crc kubenswrapper[4733]: I1204 19:45:40.094976 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck2l5" event={"ID":"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee","Type":"ContainerStarted","Data":"2be64a0f702efcf3713b3c2900f84ebd1d4a7d5bfaf17b4b2d397aed2c30db50"} Dec 04 19:45:40 crc kubenswrapper[4733]: I1204 19:45:40.141244 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ck2l5" podStartSLOduration=2.606740865 podStartE2EDuration="6.141223518s" podCreationTimestamp="2025-12-04 19:45:34 +0000 UTC" firstStartedPulling="2025-12-04 19:45:36.043083015 +0000 UTC m=+7597.998444081" lastFinishedPulling="2025-12-04 19:45:39.577565678 +0000 UTC m=+7601.532926734" observedRunningTime="2025-12-04 19:45:40.133119209 +0000 UTC m=+7602.088480255" watchObservedRunningTime="2025-12-04 19:45:40.141223518 +0000 UTC m=+7602.096584564" Dec 04 19:45:40 crc kubenswrapper[4733]: I1204 19:45:40.170924 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tvh6f"] Dec 04 19:45:41 crc kubenswrapper[4733]: I1204 19:45:41.121086 4733 generic.go:334] "Generic (PLEG): container finished" podID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerID="6ac8d28501c04023040ed0ae5abc820413631c7e9a09c246074a8f96d13880a4" exitCode=0 Dec 04 19:45:41 crc kubenswrapper[4733]: I1204 19:45:41.121190 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tvh6f" event={"ID":"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6","Type":"ContainerDied","Data":"6ac8d28501c04023040ed0ae5abc820413631c7e9a09c246074a8f96d13880a4"} Dec 04 19:45:41 crc kubenswrapper[4733]: I1204 19:45:41.122124 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tvh6f" event={"ID":"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6","Type":"ContainerStarted","Data":"d4492a43538796c1c099fbf67730f01d7c212b3918810d4468056beee0dd6abd"} Dec 04 19:45:41 crc kubenswrapper[4733]: I1204 19:45:41.124473 4733 generic.go:334] "Generic (PLEG): container finished" podID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerID="97d73bd59b14b21a92499f91c1549c5920a3d108b907c38262665b977c111135" exitCode=0 Dec 04 19:45:41 crc kubenswrapper[4733]: I1204 19:45:41.124522 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68b76" event={"ID":"ada6064a-861f-448c-9142-a3c1e6fc0f02","Type":"ContainerDied","Data":"97d73bd59b14b21a92499f91c1549c5920a3d108b907c38262665b977c111135"} Dec 04 19:45:42 crc kubenswrapper[4733]: I1204 19:45:42.136397 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68b76" event={"ID":"ada6064a-861f-448c-9142-a3c1e6fc0f02","Type":"ContainerStarted","Data":"4e4343f8ec71c7fe483e2f8400a08b3b0763f08e8f0bb8e19ab8b50f029e36b8"} Dec 04 19:45:42 crc kubenswrapper[4733]: I1204 19:45:42.167483 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-68b76" podStartSLOduration=2.515828066 podStartE2EDuration="5.167463121s" podCreationTimestamp="2025-12-04 19:45:37 +0000 UTC" firstStartedPulling="2025-12-04 19:45:39.08301551 +0000 UTC m=+7601.038376556" lastFinishedPulling="2025-12-04 19:45:41.734650565 +0000 UTC m=+7603.690011611" observedRunningTime="2025-12-04 19:45:42.155207249 +0000 UTC m=+7604.110568295" watchObservedRunningTime="2025-12-04 19:45:42.167463121 +0000 UTC m=+7604.122824167" Dec 04 19:45:44 crc kubenswrapper[4733]: I1204 19:45:44.706451 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:44 crc kubenswrapper[4733]: I1204 19:45:44.707112 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:44 crc kubenswrapper[4733]: I1204 19:45:44.761314 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:45 crc kubenswrapper[4733]: I1204 19:45:45.214619 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:45 crc kubenswrapper[4733]: I1204 19:45:45.361718 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:45:45 crc kubenswrapper[4733]: I1204 19:45:45.362026 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:45:48 crc kubenswrapper[4733]: I1204 19:45:48.113022 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:48 crc kubenswrapper[4733]: I1204 19:45:48.113500 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:48 crc kubenswrapper[4733]: I1204 19:45:48.162441 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:48 crc kubenswrapper[4733]: I1204 19:45:48.254336 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:48 crc kubenswrapper[4733]: I1204 19:45:48.363661 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ck2l5"] Dec 04 19:45:48 crc kubenswrapper[4733]: I1204 19:45:48.363875 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ck2l5" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerName="registry-server" containerID="cri-o://2be64a0f702efcf3713b3c2900f84ebd1d4a7d5bfaf17b4b2d397aed2c30db50" gracePeriod=2 Dec 04 19:45:49 crc kubenswrapper[4733]: I1204 19:45:49.220226 4733 generic.go:334] "Generic (PLEG): container finished" podID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerID="2be64a0f702efcf3713b3c2900f84ebd1d4a7d5bfaf17b4b2d397aed2c30db50" exitCode=0 Dec 04 19:45:49 crc kubenswrapper[4733]: I1204 19:45:49.220307 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck2l5" event={"ID":"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee","Type":"ContainerDied","Data":"2be64a0f702efcf3713b3c2900f84ebd1d4a7d5bfaf17b4b2d397aed2c30db50"} Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.360036 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.450219 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8j8v\" (UniqueName: \"kubernetes.io/projected/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-kube-api-access-l8j8v\") pod \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.450342 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-utilities\") pod \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.451039 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-catalog-content\") pod \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\" (UID: \"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee\") " Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.451253 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-utilities" (OuterVolumeSpecName: "utilities") pod "69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" (UID: "69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.452591 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.457085 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-kube-api-access-l8j8v" (OuterVolumeSpecName: "kube-api-access-l8j8v") pod "69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" (UID: "69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee"). InnerVolumeSpecName "kube-api-access-l8j8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.524063 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" (UID: "69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.554843 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8j8v\" (UniqueName: \"kubernetes.io/projected/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-kube-api-access-l8j8v\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:50 crc kubenswrapper[4733]: I1204 19:45:50.555108 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.239341 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tvh6f" event={"ID":"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6","Type":"ContainerStarted","Data":"0742b46534d0e8c24ccc96e3490125aa7d0f21d22bdef3fe2321538755076054"} Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.242309 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck2l5" event={"ID":"69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee","Type":"ContainerDied","Data":"7741e9818892840694acef995d66d84577e992bfea810f1faa53be2e439435d8"} Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.242418 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ck2l5" Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.242449 4733 scope.go:117] "RemoveContainer" containerID="2be64a0f702efcf3713b3c2900f84ebd1d4a7d5bfaf17b4b2d397aed2c30db50" Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.280360 4733 scope.go:117] "RemoveContainer" containerID="d4e91f2b86ea6a0ae45c88e9e59bf45a13d2718931097b55897034cf9a4bf140" Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.291907 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ck2l5"] Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.306079 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ck2l5"] Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.309929 4733 scope.go:117] "RemoveContainer" containerID="4d030d57599e004aba4d8c10008c4ce7d34161e0a816576b21cc4b45715764e6" Dec 04 19:45:51 crc kubenswrapper[4733]: E1204 19:45:51.322809 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69d03e92_bc27_4ceb_b19e_e24ca3ecf1ee.slice/crio-7741e9818892840694acef995d66d84577e992bfea810f1faa53be2e439435d8\": RecentStats: unable to find data in memory cache]" Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.764437 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68b76"] Dec 04 19:45:51 crc kubenswrapper[4733]: I1204 19:45:51.765224 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-68b76" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerName="registry-server" containerID="cri-o://4e4343f8ec71c7fe483e2f8400a08b3b0763f08e8f0bb8e19ab8b50f029e36b8" gracePeriod=2 Dec 04 19:45:52 crc kubenswrapper[4733]: I1204 19:45:52.350440 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" path="/var/lib/kubelet/pods/69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee/volumes" Dec 04 19:45:53 crc kubenswrapper[4733]: I1204 19:45:53.285775 4733 generic.go:334] "Generic (PLEG): container finished" podID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerID="4e4343f8ec71c7fe483e2f8400a08b3b0763f08e8f0bb8e19ab8b50f029e36b8" exitCode=0 Dec 04 19:45:53 crc kubenswrapper[4733]: I1204 19:45:53.285833 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68b76" event={"ID":"ada6064a-861f-448c-9142-a3c1e6fc0f02","Type":"ContainerDied","Data":"4e4343f8ec71c7fe483e2f8400a08b3b0763f08e8f0bb8e19ab8b50f029e36b8"} Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.156101 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.256745 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrgdb\" (UniqueName: \"kubernetes.io/projected/ada6064a-861f-448c-9142-a3c1e6fc0f02-kube-api-access-jrgdb\") pod \"ada6064a-861f-448c-9142-a3c1e6fc0f02\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.256984 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-utilities\") pod \"ada6064a-861f-448c-9142-a3c1e6fc0f02\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.257117 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-catalog-content\") pod \"ada6064a-861f-448c-9142-a3c1e6fc0f02\" (UID: \"ada6064a-861f-448c-9142-a3c1e6fc0f02\") " Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.257992 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-utilities" (OuterVolumeSpecName: "utilities") pod "ada6064a-861f-448c-9142-a3c1e6fc0f02" (UID: "ada6064a-861f-448c-9142-a3c1e6fc0f02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.263006 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ada6064a-861f-448c-9142-a3c1e6fc0f02-kube-api-access-jrgdb" (OuterVolumeSpecName: "kube-api-access-jrgdb") pod "ada6064a-861f-448c-9142-a3c1e6fc0f02" (UID: "ada6064a-861f-448c-9142-a3c1e6fc0f02"). InnerVolumeSpecName "kube-api-access-jrgdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.313554 4733 generic.go:334] "Generic (PLEG): container finished" podID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerID="0742b46534d0e8c24ccc96e3490125aa7d0f21d22bdef3fe2321538755076054" exitCode=0 Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.313659 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tvh6f" event={"ID":"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6","Type":"ContainerDied","Data":"0742b46534d0e8c24ccc96e3490125aa7d0f21d22bdef3fe2321538755076054"} Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.317431 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68b76" event={"ID":"ada6064a-861f-448c-9142-a3c1e6fc0f02","Type":"ContainerDied","Data":"bde8dfc193067c585fc92e9effc1f758dd9d1efdbd8941f4857871797d590153"} Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.317483 4733 scope.go:117] "RemoveContainer" containerID="4e4343f8ec71c7fe483e2f8400a08b3b0763f08e8f0bb8e19ab8b50f029e36b8" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.317866 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68b76" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.359343 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrgdb\" (UniqueName: \"kubernetes.io/projected/ada6064a-861f-448c-9142-a3c1e6fc0f02-kube-api-access-jrgdb\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.359374 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.460809 4733 scope.go:117] "RemoveContainer" containerID="97d73bd59b14b21a92499f91c1549c5920a3d108b907c38262665b977c111135" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.559835 4733 scope.go:117] "RemoveContainer" containerID="e30f0ed8b26c67cd64fff15a972d0b369934e7fd3e6b3185bc4b01e9ca63ae92" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.729180 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ada6064a-861f-448c-9142-a3c1e6fc0f02" (UID: "ada6064a-861f-448c-9142-a3c1e6fc0f02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.767239 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada6064a-861f-448c-9142-a3c1e6fc0f02-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.964231 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68b76"] Dec 04 19:45:55 crc kubenswrapper[4733]: I1204 19:45:55.977385 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-68b76"] Dec 04 19:45:56 crc kubenswrapper[4733]: I1204 19:45:56.328889 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tvh6f" event={"ID":"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6","Type":"ContainerStarted","Data":"c791df1d01ca38a8be60d594f44123326c3688546f6097125c45a4ce83fee33e"} Dec 04 19:45:56 crc kubenswrapper[4733]: I1204 19:45:56.348670 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" path="/var/lib/kubelet/pods/ada6064a-861f-448c-9142-a3c1e6fc0f02/volumes" Dec 04 19:45:56 crc kubenswrapper[4733]: I1204 19:45:56.356702 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tvh6f" podStartSLOduration=2.600280766 podStartE2EDuration="17.356676483s" podCreationTimestamp="2025-12-04 19:45:39 +0000 UTC" firstStartedPulling="2025-12-04 19:45:41.124428934 +0000 UTC m=+7603.079789970" lastFinishedPulling="2025-12-04 19:45:55.880824641 +0000 UTC m=+7617.836185687" observedRunningTime="2025-12-04 19:45:56.348107461 +0000 UTC m=+7618.303468517" watchObservedRunningTime="2025-12-04 19:45:56.356676483 +0000 UTC m=+7618.312037529" Dec 04 19:45:59 crc kubenswrapper[4733]: I1204 19:45:59.695992 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:45:59 crc kubenswrapper[4733]: I1204 19:45:59.696343 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:46:00 crc kubenswrapper[4733]: I1204 19:46:00.746652 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tvh6f" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="registry-server" probeResult="failure" output=< Dec 04 19:46:00 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 19:46:00 crc kubenswrapper[4733]: > Dec 04 19:46:09 crc kubenswrapper[4733]: I1204 19:46:09.780392 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:46:09 crc kubenswrapper[4733]: I1204 19:46:09.834557 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 19:46:10 crc kubenswrapper[4733]: I1204 19:46:10.389207 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tvh6f"] Dec 04 19:46:10 crc kubenswrapper[4733]: I1204 19:46:10.576978 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gsrhd"] Dec 04 19:46:10 crc kubenswrapper[4733]: I1204 19:46:10.577285 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gsrhd" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" containerName="registry-server" containerID="cri-o://f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea" gracePeriod=2 Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.152408 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.328461 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-utilities\") pod \"faefff68-5d9b-4191-929e-937b7b2f057e\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.328684 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-catalog-content\") pod \"faefff68-5d9b-4191-929e-937b7b2f057e\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.328713 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vpwp\" (UniqueName: \"kubernetes.io/projected/faefff68-5d9b-4191-929e-937b7b2f057e-kube-api-access-7vpwp\") pod \"faefff68-5d9b-4191-929e-937b7b2f057e\" (UID: \"faefff68-5d9b-4191-929e-937b7b2f057e\") " Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.329197 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-utilities" (OuterVolumeSpecName: "utilities") pod "faefff68-5d9b-4191-929e-937b7b2f057e" (UID: "faefff68-5d9b-4191-929e-937b7b2f057e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.330096 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.334934 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faefff68-5d9b-4191-929e-937b7b2f057e-kube-api-access-7vpwp" (OuterVolumeSpecName: "kube-api-access-7vpwp") pod "faefff68-5d9b-4191-929e-937b7b2f057e" (UID: "faefff68-5d9b-4191-929e-937b7b2f057e"). InnerVolumeSpecName "kube-api-access-7vpwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.428437 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "faefff68-5d9b-4191-929e-937b7b2f057e" (UID: "faefff68-5d9b-4191-929e-937b7b2f057e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.432812 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faefff68-5d9b-4191-929e-937b7b2f057e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.432858 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vpwp\" (UniqueName: \"kubernetes.io/projected/faefff68-5d9b-4191-929e-937b7b2f057e-kube-api-access-7vpwp\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.493214 4733 generic.go:334] "Generic (PLEG): container finished" podID="faefff68-5d9b-4191-929e-937b7b2f057e" containerID="f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea" exitCode=0 Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.493568 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gsrhd" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.493406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsrhd" event={"ID":"faefff68-5d9b-4191-929e-937b7b2f057e","Type":"ContainerDied","Data":"f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea"} Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.494179 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsrhd" event={"ID":"faefff68-5d9b-4191-929e-937b7b2f057e","Type":"ContainerDied","Data":"2544822a662e1b738db86dbdcf7fd766b47c1a0f38fa7d69feac59913a4d3f47"} Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.494346 4733 scope.go:117] "RemoveContainer" containerID="f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.524990 4733 scope.go:117] "RemoveContainer" containerID="021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.554689 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gsrhd"] Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.577244 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gsrhd"] Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.581965 4733 scope.go:117] "RemoveContainer" containerID="c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.618510 4733 scope.go:117] "RemoveContainer" containerID="f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea" Dec 04 19:46:11 crc kubenswrapper[4733]: E1204 19:46:11.619168 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea\": container with ID starting with f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea not found: ID does not exist" containerID="f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.619201 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea"} err="failed to get container status \"f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea\": rpc error: code = NotFound desc = could not find container \"f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea\": container with ID starting with f22805d1fd6c28a222b8c66fe20ae4dbca7b8070f965dbae9518099463f0c4ea not found: ID does not exist" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.619222 4733 scope.go:117] "RemoveContainer" containerID="021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3" Dec 04 19:46:11 crc kubenswrapper[4733]: E1204 19:46:11.619475 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3\": container with ID starting with 021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3 not found: ID does not exist" containerID="021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.619497 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3"} err="failed to get container status \"021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3\": rpc error: code = NotFound desc = could not find container \"021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3\": container with ID starting with 021b1303b3d06baeb588ba3e82bf72172c4086f0c83eb8d442143b5157303bf3 not found: ID does not exist" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.619509 4733 scope.go:117] "RemoveContainer" containerID="c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b" Dec 04 19:46:11 crc kubenswrapper[4733]: E1204 19:46:11.619709 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b\": container with ID starting with c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b not found: ID does not exist" containerID="c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b" Dec 04 19:46:11 crc kubenswrapper[4733]: I1204 19:46:11.619735 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b"} err="failed to get container status \"c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b\": rpc error: code = NotFound desc = could not find container \"c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b\": container with ID starting with c37dc56f2ab638ce0b4f193431ed6e665e449f5f4488dbef89a6aeac6f74423b not found: ID does not exist" Dec 04 19:46:12 crc kubenswrapper[4733]: I1204 19:46:12.348554 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" path="/var/lib/kubelet/pods/faefff68-5d9b-4191-929e-937b7b2f057e/volumes" Dec 04 19:46:15 crc kubenswrapper[4733]: I1204 19:46:15.362541 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:46:15 crc kubenswrapper[4733]: I1204 19:46:15.362944 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:46:37 crc kubenswrapper[4733]: I1204 19:46:37.806964 4733 generic.go:334] "Generic (PLEG): container finished" podID="1b7b4789-500e-40d1-bddd-332ab22536e9" containerID="613a370d675ce7c1d82cb5b42d705ff072705cf2c8586f051107e1d86bb746c3" exitCode=0 Dec 04 19:46:37 crc kubenswrapper[4733]: I1204 19:46:37.807083 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-9v588" event={"ID":"1b7b4789-500e-40d1-bddd-332ab22536e9","Type":"ContainerDied","Data":"613a370d675ce7c1d82cb5b42d705ff072705cf2c8586f051107e1d86bb746c3"} Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.406562 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.503373 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prh9h\" (UniqueName: \"kubernetes.io/projected/1b7b4789-500e-40d1-bddd-332ab22536e9-kube-api-access-prh9h\") pod \"1b7b4789-500e-40d1-bddd-332ab22536e9\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.503438 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ceph\") pod \"1b7b4789-500e-40d1-bddd-332ab22536e9\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.503478 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ssh-key\") pod \"1b7b4789-500e-40d1-bddd-332ab22536e9\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.503502 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-inventory\") pod \"1b7b4789-500e-40d1-bddd-332ab22536e9\" (UID: \"1b7b4789-500e-40d1-bddd-332ab22536e9\") " Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.508978 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ceph" (OuterVolumeSpecName: "ceph") pod "1b7b4789-500e-40d1-bddd-332ab22536e9" (UID: "1b7b4789-500e-40d1-bddd-332ab22536e9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.510269 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b7b4789-500e-40d1-bddd-332ab22536e9-kube-api-access-prh9h" (OuterVolumeSpecName: "kube-api-access-prh9h") pod "1b7b4789-500e-40d1-bddd-332ab22536e9" (UID: "1b7b4789-500e-40d1-bddd-332ab22536e9"). InnerVolumeSpecName "kube-api-access-prh9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.537575 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-inventory" (OuterVolumeSpecName: "inventory") pod "1b7b4789-500e-40d1-bddd-332ab22536e9" (UID: "1b7b4789-500e-40d1-bddd-332ab22536e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.545838 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1b7b4789-500e-40d1-bddd-332ab22536e9" (UID: "1b7b4789-500e-40d1-bddd-332ab22536e9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.605512 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prh9h\" (UniqueName: \"kubernetes.io/projected/1b7b4789-500e-40d1-bddd-332ab22536e9-kube-api-access-prh9h\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.605561 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.605577 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.605590 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b7b4789-500e-40d1-bddd-332ab22536e9-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.834084 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-9v588" event={"ID":"1b7b4789-500e-40d1-bddd-332ab22536e9","Type":"ContainerDied","Data":"74544c3e3b76ed7240b38b73eb37a206687cbf4b0ba4cb76d597b7b91075b160"} Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.834464 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74544c3e3b76ed7240b38b73eb37a206687cbf4b0ba4cb76d597b7b91075b160" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.834148 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-9v588" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.914868 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-s5fzv"] Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915394 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915417 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915434 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerName="extract-utilities" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915444 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerName="extract-utilities" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915464 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerName="extract-content" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915472 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerName="extract-content" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915502 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" containerName="extract-content" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915511 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" containerName="extract-content" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915520 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915529 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915551 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerName="extract-content" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915559 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerName="extract-content" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915578 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b7b4789-500e-40d1-bddd-332ab22536e9" containerName="configure-network-openstack-openstack-cell1" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915588 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b7b4789-500e-40d1-bddd-332ab22536e9" containerName="configure-network-openstack-openstack-cell1" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915607 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerName="extract-utilities" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915616 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerName="extract-utilities" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915650 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" containerName="extract-utilities" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915658 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" containerName="extract-utilities" Dec 04 19:46:39 crc kubenswrapper[4733]: E1204 19:46:39.915672 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915682 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915948 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ada6064a-861f-448c-9142-a3c1e6fc0f02" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915970 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b7b4789-500e-40d1-bddd-332ab22536e9" containerName="configure-network-openstack-openstack-cell1" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.915992 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="faefff68-5d9b-4191-929e-937b7b2f057e" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.916019 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="69d03e92-bc27-4ceb-b19e-e24ca3ecf1ee" containerName="registry-server" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.916992 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.919646 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.920329 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.920888 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.921275 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:46:39 crc kubenswrapper[4733]: I1204 19:46:39.934849 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-s5fzv"] Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.015173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-inventory\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.015227 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ceph\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.015331 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ssh-key\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.015970 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbqql\" (UniqueName: \"kubernetes.io/projected/32a595da-5163-4f68-99f5-195b933e51f9-kube-api-access-kbqql\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.117367 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-inventory\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.117423 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ceph\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.117670 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ssh-key\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.118023 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbqql\" (UniqueName: \"kubernetes.io/projected/32a595da-5163-4f68-99f5-195b933e51f9-kube-api-access-kbqql\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.122526 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-inventory\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.122582 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ceph\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.122943 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ssh-key\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.137968 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbqql\" (UniqueName: \"kubernetes.io/projected/32a595da-5163-4f68-99f5-195b933e51f9-kube-api-access-kbqql\") pod \"validate-network-openstack-openstack-cell1-s5fzv\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.254132 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.836775 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-s5fzv"] Dec 04 19:46:40 crc kubenswrapper[4733]: I1204 19:46:40.843598 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:46:41 crc kubenswrapper[4733]: I1204 19:46:41.858274 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" event={"ID":"32a595da-5163-4f68-99f5-195b933e51f9","Type":"ContainerStarted","Data":"93db6b13e4a0982abdab69e7f822f43667fb53a63de038ce85e7144e1400fc82"} Dec 04 19:46:41 crc kubenswrapper[4733]: I1204 19:46:41.858577 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" event={"ID":"32a595da-5163-4f68-99f5-195b933e51f9","Type":"ContainerStarted","Data":"c1b2efea717d349c8853f585094f48bf641543369160b1239cfdb71014a7ee5f"} Dec 04 19:46:41 crc kubenswrapper[4733]: I1204 19:46:41.884187 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" podStartSLOduration=2.369651122 podStartE2EDuration="2.884162871s" podCreationTimestamp="2025-12-04 19:46:39 +0000 UTC" firstStartedPulling="2025-12-04 19:46:40.843382035 +0000 UTC m=+7662.798743081" lastFinishedPulling="2025-12-04 19:46:41.357893784 +0000 UTC m=+7663.313254830" observedRunningTime="2025-12-04 19:46:41.872350601 +0000 UTC m=+7663.827711657" watchObservedRunningTime="2025-12-04 19:46:41.884162871 +0000 UTC m=+7663.839523937" Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.362414 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.363169 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.363238 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.364836 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6702c9ff250314030e3d19e8ac891392736fe3f613350f0fa8b792973e875133"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.364944 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://6702c9ff250314030e3d19e8ac891392736fe3f613350f0fa8b792973e875133" gracePeriod=600 Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.901971 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="6702c9ff250314030e3d19e8ac891392736fe3f613350f0fa8b792973e875133" exitCode=0 Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.902055 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"6702c9ff250314030e3d19e8ac891392736fe3f613350f0fa8b792973e875133"} Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.902625 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d"} Dec 04 19:46:45 crc kubenswrapper[4733]: I1204 19:46:45.902661 4733 scope.go:117] "RemoveContainer" containerID="3e8de39fcfb692051f8c7781d5583bd04bb7d0656ebe6fffb8466d6c57e71b96" Dec 04 19:46:46 crc kubenswrapper[4733]: I1204 19:46:46.913194 4733 generic.go:334] "Generic (PLEG): container finished" podID="32a595da-5163-4f68-99f5-195b933e51f9" containerID="93db6b13e4a0982abdab69e7f822f43667fb53a63de038ce85e7144e1400fc82" exitCode=0 Dec 04 19:46:46 crc kubenswrapper[4733]: I1204 19:46:46.913351 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" event={"ID":"32a595da-5163-4f68-99f5-195b933e51f9","Type":"ContainerDied","Data":"93db6b13e4a0982abdab69e7f822f43667fb53a63de038ce85e7144e1400fc82"} Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.504726 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.603528 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbqql\" (UniqueName: \"kubernetes.io/projected/32a595da-5163-4f68-99f5-195b933e51f9-kube-api-access-kbqql\") pod \"32a595da-5163-4f68-99f5-195b933e51f9\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.603640 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-inventory\") pod \"32a595da-5163-4f68-99f5-195b933e51f9\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.603731 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ceph\") pod \"32a595da-5163-4f68-99f5-195b933e51f9\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.603890 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ssh-key\") pod \"32a595da-5163-4f68-99f5-195b933e51f9\" (UID: \"32a595da-5163-4f68-99f5-195b933e51f9\") " Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.609416 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ceph" (OuterVolumeSpecName: "ceph") pod "32a595da-5163-4f68-99f5-195b933e51f9" (UID: "32a595da-5163-4f68-99f5-195b933e51f9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.609677 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a595da-5163-4f68-99f5-195b933e51f9-kube-api-access-kbqql" (OuterVolumeSpecName: "kube-api-access-kbqql") pod "32a595da-5163-4f68-99f5-195b933e51f9" (UID: "32a595da-5163-4f68-99f5-195b933e51f9"). InnerVolumeSpecName "kube-api-access-kbqql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.635542 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-inventory" (OuterVolumeSpecName: "inventory") pod "32a595da-5163-4f68-99f5-195b933e51f9" (UID: "32a595da-5163-4f68-99f5-195b933e51f9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.643738 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "32a595da-5163-4f68-99f5-195b933e51f9" (UID: "32a595da-5163-4f68-99f5-195b933e51f9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.707087 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.707128 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbqql\" (UniqueName: \"kubernetes.io/projected/32a595da-5163-4f68-99f5-195b933e51f9-kube-api-access-kbqql\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.707173 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.707185 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/32a595da-5163-4f68-99f5-195b933e51f9-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.940488 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" event={"ID":"32a595da-5163-4f68-99f5-195b933e51f9","Type":"ContainerDied","Data":"c1b2efea717d349c8853f585094f48bf641543369160b1239cfdb71014a7ee5f"} Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.940862 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1b2efea717d349c8853f585094f48bf641543369160b1239cfdb71014a7ee5f" Dec 04 19:46:48 crc kubenswrapper[4733]: I1204 19:46:48.940684 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-s5fzv" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.051282 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-fpctx"] Dec 04 19:46:49 crc kubenswrapper[4733]: E1204 19:46:49.051757 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a595da-5163-4f68-99f5-195b933e51f9" containerName="validate-network-openstack-openstack-cell1" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.051779 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a595da-5163-4f68-99f5-195b933e51f9" containerName="validate-network-openstack-openstack-cell1" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.052090 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a595da-5163-4f68-99f5-195b933e51f9" containerName="validate-network-openstack-openstack-cell1" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.052961 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.055452 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.056344 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.056533 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.058510 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.069461 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-fpctx"] Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.114265 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ssh-key\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.114325 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snrnn\" (UniqueName: \"kubernetes.io/projected/0899052c-379c-4fd4-869b-b49d488cc60b-kube-api-access-snrnn\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.114351 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ceph\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.114390 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-inventory\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.216083 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snrnn\" (UniqueName: \"kubernetes.io/projected/0899052c-379c-4fd4-869b-b49d488cc60b-kube-api-access-snrnn\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.216182 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ceph\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.216271 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-inventory\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.216534 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ssh-key\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.222061 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-inventory\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.223101 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ssh-key\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.224043 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ceph\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.246020 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snrnn\" (UniqueName: \"kubernetes.io/projected/0899052c-379c-4fd4-869b-b49d488cc60b-kube-api-access-snrnn\") pod \"install-os-openstack-openstack-cell1-fpctx\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.412870 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:46:49 crc kubenswrapper[4733]: I1204 19:46:49.985235 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-fpctx"] Dec 04 19:46:50 crc kubenswrapper[4733]: I1204 19:46:50.964017 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-fpctx" event={"ID":"0899052c-379c-4fd4-869b-b49d488cc60b","Type":"ContainerStarted","Data":"1462ebf0c6c20238c59ea363dae510fcedb21503c41f14144cc35159ed5faa14"} Dec 04 19:46:50 crc kubenswrapper[4733]: I1204 19:46:50.964445 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-fpctx" event={"ID":"0899052c-379c-4fd4-869b-b49d488cc60b","Type":"ContainerStarted","Data":"b1bbf680a750ab65cdfa05c2d05493aef7c9d5e702ce8ad8acf3d8049ee37e8d"} Dec 04 19:46:50 crc kubenswrapper[4733]: I1204 19:46:50.992439 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-fpctx" podStartSLOduration=1.5285800680000001 podStartE2EDuration="1.992411543s" podCreationTimestamp="2025-12-04 19:46:49 +0000 UTC" firstStartedPulling="2025-12-04 19:46:50.00561813 +0000 UTC m=+7671.960979176" lastFinishedPulling="2025-12-04 19:46:50.469449595 +0000 UTC m=+7672.424810651" observedRunningTime="2025-12-04 19:46:50.978560717 +0000 UTC m=+7672.933921773" watchObservedRunningTime="2025-12-04 19:46:50.992411543 +0000 UTC m=+7672.947772609" Dec 04 19:47:28 crc kubenswrapper[4733]: I1204 19:47:28.395740 4733 scope.go:117] "RemoveContainer" containerID="26439a8495ca2993ce216c8c8abf6136904fc14aac585b1dfc5b4c09521fa54e" Dec 04 19:47:28 crc kubenswrapper[4733]: I1204 19:47:28.436544 4733 scope.go:117] "RemoveContainer" containerID="fda5997a954620b5f0eef7bd8da730562e3f25a363a2d5ae1ee25e4f75715242" Dec 04 19:47:28 crc kubenswrapper[4733]: I1204 19:47:28.496325 4733 scope.go:117] "RemoveContainer" containerID="72aaddc5b841f7123b30a7d619522d22916424c1f803ac29e6e82d907d94cece" Dec 04 19:47:36 crc kubenswrapper[4733]: I1204 19:47:36.525901 4733 generic.go:334] "Generic (PLEG): container finished" podID="0899052c-379c-4fd4-869b-b49d488cc60b" containerID="1462ebf0c6c20238c59ea363dae510fcedb21503c41f14144cc35159ed5faa14" exitCode=0 Dec 04 19:47:36 crc kubenswrapper[4733]: I1204 19:47:36.526012 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-fpctx" event={"ID":"0899052c-379c-4fd4-869b-b49d488cc60b","Type":"ContainerDied","Data":"1462ebf0c6c20238c59ea363dae510fcedb21503c41f14144cc35159ed5faa14"} Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.014242 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.102510 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ceph\") pod \"0899052c-379c-4fd4-869b-b49d488cc60b\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.103467 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ssh-key\") pod \"0899052c-379c-4fd4-869b-b49d488cc60b\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.103536 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-inventory\") pod \"0899052c-379c-4fd4-869b-b49d488cc60b\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.103582 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snrnn\" (UniqueName: \"kubernetes.io/projected/0899052c-379c-4fd4-869b-b49d488cc60b-kube-api-access-snrnn\") pod \"0899052c-379c-4fd4-869b-b49d488cc60b\" (UID: \"0899052c-379c-4fd4-869b-b49d488cc60b\") " Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.111685 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0899052c-379c-4fd4-869b-b49d488cc60b-kube-api-access-snrnn" (OuterVolumeSpecName: "kube-api-access-snrnn") pod "0899052c-379c-4fd4-869b-b49d488cc60b" (UID: "0899052c-379c-4fd4-869b-b49d488cc60b"). InnerVolumeSpecName "kube-api-access-snrnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.119196 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ceph" (OuterVolumeSpecName: "ceph") pod "0899052c-379c-4fd4-869b-b49d488cc60b" (UID: "0899052c-379c-4fd4-869b-b49d488cc60b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.140427 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0899052c-379c-4fd4-869b-b49d488cc60b" (UID: "0899052c-379c-4fd4-869b-b49d488cc60b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.150616 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-inventory" (OuterVolumeSpecName: "inventory") pod "0899052c-379c-4fd4-869b-b49d488cc60b" (UID: "0899052c-379c-4fd4-869b-b49d488cc60b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.206271 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.206321 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.206334 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0899052c-379c-4fd4-869b-b49d488cc60b-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.206347 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snrnn\" (UniqueName: \"kubernetes.io/projected/0899052c-379c-4fd4-869b-b49d488cc60b-kube-api-access-snrnn\") on node \"crc\" DevicePath \"\"" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.550231 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-fpctx" event={"ID":"0899052c-379c-4fd4-869b-b49d488cc60b","Type":"ContainerDied","Data":"b1bbf680a750ab65cdfa05c2d05493aef7c9d5e702ce8ad8acf3d8049ee37e8d"} Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.550527 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1bbf680a750ab65cdfa05c2d05493aef7c9d5e702ce8ad8acf3d8049ee37e8d" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.550345 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-fpctx" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.648777 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-l82wt"] Dec 04 19:47:38 crc kubenswrapper[4733]: E1204 19:47:38.649318 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0899052c-379c-4fd4-869b-b49d488cc60b" containerName="install-os-openstack-openstack-cell1" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.649342 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0899052c-379c-4fd4-869b-b49d488cc60b" containerName="install-os-openstack-openstack-cell1" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.649608 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0899052c-379c-4fd4-869b-b49d488cc60b" containerName="install-os-openstack-openstack-cell1" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.650604 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.653598 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.653934 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.654496 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.654775 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.669135 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-l82wt"] Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.822869 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ssh-key\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.823054 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ceph\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.823105 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-inventory\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.823256 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54zkx\" (UniqueName: \"kubernetes.io/projected/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-kube-api-access-54zkx\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.925483 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54zkx\" (UniqueName: \"kubernetes.io/projected/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-kube-api-access-54zkx\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.925644 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ssh-key\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.925883 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ceph\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.925950 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-inventory\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.931353 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ceph\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.931406 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-inventory\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.935332 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ssh-key\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.955606 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54zkx\" (UniqueName: \"kubernetes.io/projected/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-kube-api-access-54zkx\") pod \"configure-os-openstack-openstack-cell1-l82wt\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:38 crc kubenswrapper[4733]: I1204 19:47:38.974559 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:47:39 crc kubenswrapper[4733]: I1204 19:47:39.519606 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-l82wt"] Dec 04 19:47:39 crc kubenswrapper[4733]: I1204 19:47:39.563455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" event={"ID":"27641fdb-1ce8-44d2-b0f6-f0b8cc303338","Type":"ContainerStarted","Data":"db05aac1c5b0425d3e9d4e5ee92af422b5af069541a2ea18377229ebc54b7023"} Dec 04 19:47:40 crc kubenswrapper[4733]: I1204 19:47:40.578364 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" event={"ID":"27641fdb-1ce8-44d2-b0f6-f0b8cc303338","Type":"ContainerStarted","Data":"bd5e6adffc7838380fb2daba09cbb1e96905db106a53d90632910d8805ea2f6b"} Dec 04 19:47:40 crc kubenswrapper[4733]: I1204 19:47:40.610194 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" podStartSLOduration=2.178207903 podStartE2EDuration="2.610170997s" podCreationTimestamp="2025-12-04 19:47:38 +0000 UTC" firstStartedPulling="2025-12-04 19:47:39.521405375 +0000 UTC m=+7721.476766441" lastFinishedPulling="2025-12-04 19:47:39.953368449 +0000 UTC m=+7721.908729535" observedRunningTime="2025-12-04 19:47:40.603317891 +0000 UTC m=+7722.558678967" watchObservedRunningTime="2025-12-04 19:47:40.610170997 +0000 UTC m=+7722.565532083" Dec 04 19:48:26 crc kubenswrapper[4733]: I1204 19:48:26.109429 4733 generic.go:334] "Generic (PLEG): container finished" podID="27641fdb-1ce8-44d2-b0f6-f0b8cc303338" containerID="bd5e6adffc7838380fb2daba09cbb1e96905db106a53d90632910d8805ea2f6b" exitCode=0 Dec 04 19:48:26 crc kubenswrapper[4733]: I1204 19:48:26.109509 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" event={"ID":"27641fdb-1ce8-44d2-b0f6-f0b8cc303338","Type":"ContainerDied","Data":"bd5e6adffc7838380fb2daba09cbb1e96905db106a53d90632910d8805ea2f6b"} Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.601873 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.725978 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-inventory\") pod \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.726146 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ssh-key\") pod \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.726546 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ceph\") pod \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.726719 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54zkx\" (UniqueName: \"kubernetes.io/projected/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-kube-api-access-54zkx\") pod \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\" (UID: \"27641fdb-1ce8-44d2-b0f6-f0b8cc303338\") " Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.733295 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ceph" (OuterVolumeSpecName: "ceph") pod "27641fdb-1ce8-44d2-b0f6-f0b8cc303338" (UID: "27641fdb-1ce8-44d2-b0f6-f0b8cc303338"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.733456 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-kube-api-access-54zkx" (OuterVolumeSpecName: "kube-api-access-54zkx") pod "27641fdb-1ce8-44d2-b0f6-f0b8cc303338" (UID: "27641fdb-1ce8-44d2-b0f6-f0b8cc303338"). InnerVolumeSpecName "kube-api-access-54zkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.761129 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-inventory" (OuterVolumeSpecName: "inventory") pod "27641fdb-1ce8-44d2-b0f6-f0b8cc303338" (UID: "27641fdb-1ce8-44d2-b0f6-f0b8cc303338"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.766780 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27641fdb-1ce8-44d2-b0f6-f0b8cc303338" (UID: "27641fdb-1ce8-44d2-b0f6-f0b8cc303338"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.829370 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.829640 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54zkx\" (UniqueName: \"kubernetes.io/projected/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-kube-api-access-54zkx\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.829757 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:27 crc kubenswrapper[4733]: I1204 19:48:27.829898 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27641fdb-1ce8-44d2-b0f6-f0b8cc303338-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.132476 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" event={"ID":"27641fdb-1ce8-44d2-b0f6-f0b8cc303338","Type":"ContainerDied","Data":"db05aac1c5b0425d3e9d4e5ee92af422b5af069541a2ea18377229ebc54b7023"} Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.132520 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db05aac1c5b0425d3e9d4e5ee92af422b5af069541a2ea18377229ebc54b7023" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.132579 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-l82wt" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.220265 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-tnzhn"] Dec 04 19:48:28 crc kubenswrapper[4733]: E1204 19:48:28.220712 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27641fdb-1ce8-44d2-b0f6-f0b8cc303338" containerName="configure-os-openstack-openstack-cell1" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.220730 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="27641fdb-1ce8-44d2-b0f6-f0b8cc303338" containerName="configure-os-openstack-openstack-cell1" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.220997 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="27641fdb-1ce8-44d2-b0f6-f0b8cc303338" containerName="configure-os-openstack-openstack-cell1" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.229315 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.230153 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-tnzhn"] Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.232202 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.232848 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.232894 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.232904 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.340696 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ceph\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.341178 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-inventory-0\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.341311 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clnr7\" (UniqueName: \"kubernetes.io/projected/f485e10f-8830-4d98-ad15-ff39e157e28c-kube-api-access-clnr7\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.341621 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.443883 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.444056 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ceph\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.444312 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-inventory-0\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.444417 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clnr7\" (UniqueName: \"kubernetes.io/projected/f485e10f-8830-4d98-ad15-ff39e157e28c-kube-api-access-clnr7\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.449448 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ceph\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.449521 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-inventory-0\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.458940 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.468768 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clnr7\" (UniqueName: \"kubernetes.io/projected/f485e10f-8830-4d98-ad15-ff39e157e28c-kube-api-access-clnr7\") pod \"ssh-known-hosts-openstack-tnzhn\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:28 crc kubenswrapper[4733]: I1204 19:48:28.546956 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:29 crc kubenswrapper[4733]: I1204 19:48:29.094577 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-tnzhn"] Dec 04 19:48:29 crc kubenswrapper[4733]: I1204 19:48:29.142185 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-tnzhn" event={"ID":"f485e10f-8830-4d98-ad15-ff39e157e28c","Type":"ContainerStarted","Data":"dcdbb52075f2bb5675b73331450e86f378dc48c887402f76237fe7932219bfc2"} Dec 04 19:48:30 crc kubenswrapper[4733]: I1204 19:48:30.157339 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-tnzhn" event={"ID":"f485e10f-8830-4d98-ad15-ff39e157e28c","Type":"ContainerStarted","Data":"45995f64abb53676d7c7c47b45648bead001c6b5e44a416168461aa94a0a4310"} Dec 04 19:48:30 crc kubenswrapper[4733]: I1204 19:48:30.189189 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-tnzhn" podStartSLOduration=1.664536282 podStartE2EDuration="2.189162377s" podCreationTimestamp="2025-12-04 19:48:28 +0000 UTC" firstStartedPulling="2025-12-04 19:48:29.102864013 +0000 UTC m=+7771.058225059" lastFinishedPulling="2025-12-04 19:48:29.627490108 +0000 UTC m=+7771.582851154" observedRunningTime="2025-12-04 19:48:30.17604105 +0000 UTC m=+7772.131402136" watchObservedRunningTime="2025-12-04 19:48:30.189162377 +0000 UTC m=+7772.144523443" Dec 04 19:48:39 crc kubenswrapper[4733]: I1204 19:48:39.258851 4733 generic.go:334] "Generic (PLEG): container finished" podID="f485e10f-8830-4d98-ad15-ff39e157e28c" containerID="45995f64abb53676d7c7c47b45648bead001c6b5e44a416168461aa94a0a4310" exitCode=0 Dec 04 19:48:39 crc kubenswrapper[4733]: I1204 19:48:39.259012 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-tnzhn" event={"ID":"f485e10f-8830-4d98-ad15-ff39e157e28c","Type":"ContainerDied","Data":"45995f64abb53676d7c7c47b45648bead001c6b5e44a416168461aa94a0a4310"} Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.817616 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.929309 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clnr7\" (UniqueName: \"kubernetes.io/projected/f485e10f-8830-4d98-ad15-ff39e157e28c-kube-api-access-clnr7\") pod \"f485e10f-8830-4d98-ad15-ff39e157e28c\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.929657 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ssh-key-openstack-cell1\") pod \"f485e10f-8830-4d98-ad15-ff39e157e28c\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.929694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-inventory-0\") pod \"f485e10f-8830-4d98-ad15-ff39e157e28c\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.929747 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ceph\") pod \"f485e10f-8830-4d98-ad15-ff39e157e28c\" (UID: \"f485e10f-8830-4d98-ad15-ff39e157e28c\") " Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.937142 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f485e10f-8830-4d98-ad15-ff39e157e28c-kube-api-access-clnr7" (OuterVolumeSpecName: "kube-api-access-clnr7") pod "f485e10f-8830-4d98-ad15-ff39e157e28c" (UID: "f485e10f-8830-4d98-ad15-ff39e157e28c"). InnerVolumeSpecName "kube-api-access-clnr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.937440 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ceph" (OuterVolumeSpecName: "ceph") pod "f485e10f-8830-4d98-ad15-ff39e157e28c" (UID: "f485e10f-8830-4d98-ad15-ff39e157e28c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.961886 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "f485e10f-8830-4d98-ad15-ff39e157e28c" (UID: "f485e10f-8830-4d98-ad15-ff39e157e28c"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:40 crc kubenswrapper[4733]: I1204 19:48:40.975228 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "f485e10f-8830-4d98-ad15-ff39e157e28c" (UID: "f485e10f-8830-4d98-ad15-ff39e157e28c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.032319 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clnr7\" (UniqueName: \"kubernetes.io/projected/f485e10f-8830-4d98-ad15-ff39e157e28c-kube-api-access-clnr7\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.032370 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.032389 4733 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.032406 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f485e10f-8830-4d98-ad15-ff39e157e28c-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.292128 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-tnzhn" event={"ID":"f485e10f-8830-4d98-ad15-ff39e157e28c","Type":"ContainerDied","Data":"dcdbb52075f2bb5675b73331450e86f378dc48c887402f76237fe7932219bfc2"} Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.292174 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcdbb52075f2bb5675b73331450e86f378dc48c887402f76237fe7932219bfc2" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.292253 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-tnzhn" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.363229 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-7ddvl"] Dec 04 19:48:41 crc kubenswrapper[4733]: E1204 19:48:41.363922 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f485e10f-8830-4d98-ad15-ff39e157e28c" containerName="ssh-known-hosts-openstack" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.363946 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f485e10f-8830-4d98-ad15-ff39e157e28c" containerName="ssh-known-hosts-openstack" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.364249 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f485e10f-8830-4d98-ad15-ff39e157e28c" containerName="ssh-known-hosts-openstack" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.365273 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.371316 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.371411 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.371425 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.371851 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.394590 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-7ddvl"] Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.439104 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ceph\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.439221 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-inventory\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.439691 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxxg5\" (UniqueName: \"kubernetes.io/projected/3ac10d63-265c-49a1-a66c-bd638c729159-kube-api-access-rxxg5\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.439770 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ssh-key\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.542302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxxg5\" (UniqueName: \"kubernetes.io/projected/3ac10d63-265c-49a1-a66c-bd638c729159-kube-api-access-rxxg5\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.542366 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ssh-key\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.542420 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ceph\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.542504 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-inventory\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.547032 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-inventory\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.547454 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ceph\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.549207 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ssh-key\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.560511 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxxg5\" (UniqueName: \"kubernetes.io/projected/3ac10d63-265c-49a1-a66c-bd638c729159-kube-api-access-rxxg5\") pod \"run-os-openstack-openstack-cell1-7ddvl\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:41 crc kubenswrapper[4733]: I1204 19:48:41.693550 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:42 crc kubenswrapper[4733]: I1204 19:48:42.291034 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-7ddvl"] Dec 04 19:48:43 crc kubenswrapper[4733]: I1204 19:48:43.325300 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" event={"ID":"3ac10d63-265c-49a1-a66c-bd638c729159","Type":"ContainerStarted","Data":"fb25701a2298cc15d48151863db0941e13ab41f706f168a8de2655ed3ab3b9cb"} Dec 04 19:48:43 crc kubenswrapper[4733]: I1204 19:48:43.325982 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" event={"ID":"3ac10d63-265c-49a1-a66c-bd638c729159","Type":"ContainerStarted","Data":"0ac3f6e78256319d5583d1b4f23a07b10638e61c1100749ed786bde295f0d497"} Dec 04 19:48:43 crc kubenswrapper[4733]: I1204 19:48:43.356279 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" podStartSLOduration=1.887363294 podStartE2EDuration="2.356257987s" podCreationTimestamp="2025-12-04 19:48:41 +0000 UTC" firstStartedPulling="2025-12-04 19:48:42.298727494 +0000 UTC m=+7784.254088540" lastFinishedPulling="2025-12-04 19:48:42.767622187 +0000 UTC m=+7784.722983233" observedRunningTime="2025-12-04 19:48:43.353265146 +0000 UTC m=+7785.308626242" watchObservedRunningTime="2025-12-04 19:48:43.356257987 +0000 UTC m=+7785.311619033" Dec 04 19:48:45 crc kubenswrapper[4733]: I1204 19:48:45.361948 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:48:45 crc kubenswrapper[4733]: I1204 19:48:45.362435 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:48:51 crc kubenswrapper[4733]: I1204 19:48:51.422006 4733 generic.go:334] "Generic (PLEG): container finished" podID="3ac10d63-265c-49a1-a66c-bd638c729159" containerID="fb25701a2298cc15d48151863db0941e13ab41f706f168a8de2655ed3ab3b9cb" exitCode=0 Dec 04 19:48:51 crc kubenswrapper[4733]: I1204 19:48:51.422158 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" event={"ID":"3ac10d63-265c-49a1-a66c-bd638c729159","Type":"ContainerDied","Data":"fb25701a2298cc15d48151863db0941e13ab41f706f168a8de2655ed3ab3b9cb"} Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.891704 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.916771 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ssh-key\") pod \"3ac10d63-265c-49a1-a66c-bd638c729159\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.916910 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ceph\") pod \"3ac10d63-265c-49a1-a66c-bd638c729159\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.917040 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-inventory\") pod \"3ac10d63-265c-49a1-a66c-bd638c729159\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.917063 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxxg5\" (UniqueName: \"kubernetes.io/projected/3ac10d63-265c-49a1-a66c-bd638c729159-kube-api-access-rxxg5\") pod \"3ac10d63-265c-49a1-a66c-bd638c729159\" (UID: \"3ac10d63-265c-49a1-a66c-bd638c729159\") " Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.924484 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ceph" (OuterVolumeSpecName: "ceph") pod "3ac10d63-265c-49a1-a66c-bd638c729159" (UID: "3ac10d63-265c-49a1-a66c-bd638c729159"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.924751 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ac10d63-265c-49a1-a66c-bd638c729159-kube-api-access-rxxg5" (OuterVolumeSpecName: "kube-api-access-rxxg5") pod "3ac10d63-265c-49a1-a66c-bd638c729159" (UID: "3ac10d63-265c-49a1-a66c-bd638c729159"). InnerVolumeSpecName "kube-api-access-rxxg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.961052 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-inventory" (OuterVolumeSpecName: "inventory") pod "3ac10d63-265c-49a1-a66c-bd638c729159" (UID: "3ac10d63-265c-49a1-a66c-bd638c729159"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:52 crc kubenswrapper[4733]: I1204 19:48:52.980410 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3ac10d63-265c-49a1-a66c-bd638c729159" (UID: "3ac10d63-265c-49a1-a66c-bd638c729159"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.019019 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.019046 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.019055 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ac10d63-265c-49a1-a66c-bd638c729159-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.019064 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxxg5\" (UniqueName: \"kubernetes.io/projected/3ac10d63-265c-49a1-a66c-bd638c729159-kube-api-access-rxxg5\") on node \"crc\" DevicePath \"\"" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.448617 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" event={"ID":"3ac10d63-265c-49a1-a66c-bd638c729159","Type":"ContainerDied","Data":"0ac3f6e78256319d5583d1b4f23a07b10638e61c1100749ed786bde295f0d497"} Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.448675 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ac3f6e78256319d5583d1b4f23a07b10638e61c1100749ed786bde295f0d497" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.449271 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-7ddvl" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.534729 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-ck6bp"] Dec 04 19:48:53 crc kubenswrapper[4733]: E1204 19:48:53.537579 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac10d63-265c-49a1-a66c-bd638c729159" containerName="run-os-openstack-openstack-cell1" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.537610 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac10d63-265c-49a1-a66c-bd638c729159" containerName="run-os-openstack-openstack-cell1" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.537956 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ac10d63-265c-49a1-a66c-bd638c729159" containerName="run-os-openstack-openstack-cell1" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.539166 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.551749 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-ck6bp"] Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.577403 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.577988 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.578126 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.580572 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.631720 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.632049 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-676rn\" (UniqueName: \"kubernetes.io/projected/52516c81-59f5-4b50-9b16-80fcc8303cc9-kube-api-access-676rn\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.632089 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ceph\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.632172 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-inventory\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.733769 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-676rn\" (UniqueName: \"kubernetes.io/projected/52516c81-59f5-4b50-9b16-80fcc8303cc9-kube-api-access-676rn\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.733846 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ceph\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.733974 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-inventory\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.734018 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.741039 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.741199 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ceph\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.741561 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-inventory\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.750567 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-676rn\" (UniqueName: \"kubernetes.io/projected/52516c81-59f5-4b50-9b16-80fcc8303cc9-kube-api-access-676rn\") pod \"reboot-os-openstack-openstack-cell1-ck6bp\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:53 crc kubenswrapper[4733]: I1204 19:48:53.903572 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:48:54 crc kubenswrapper[4733]: I1204 19:48:54.461115 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-ck6bp"] Dec 04 19:48:55 crc kubenswrapper[4733]: I1204 19:48:55.479323 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" event={"ID":"52516c81-59f5-4b50-9b16-80fcc8303cc9","Type":"ContainerStarted","Data":"70ff1201f7b65cc75fa50942282185e1f5b50ca891abbf21e514ecb3921235c6"} Dec 04 19:48:55 crc kubenswrapper[4733]: I1204 19:48:55.480010 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" event={"ID":"52516c81-59f5-4b50-9b16-80fcc8303cc9","Type":"ContainerStarted","Data":"d27574df0ea0ed4c09fdaaa68d29d5944f43b9ba4ceb01bf750851bf55c0a450"} Dec 04 19:48:55 crc kubenswrapper[4733]: I1204 19:48:55.509449 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" podStartSLOduration=1.940421718 podStartE2EDuration="2.509423025s" podCreationTimestamp="2025-12-04 19:48:53 +0000 UTC" firstStartedPulling="2025-12-04 19:48:54.461926755 +0000 UTC m=+7796.417287831" lastFinishedPulling="2025-12-04 19:48:55.030928052 +0000 UTC m=+7796.986289138" observedRunningTime="2025-12-04 19:48:55.50406793 +0000 UTC m=+7797.459429046" watchObservedRunningTime="2025-12-04 19:48:55.509423025 +0000 UTC m=+7797.464784081" Dec 04 19:49:11 crc kubenswrapper[4733]: I1204 19:49:11.675242 4733 generic.go:334] "Generic (PLEG): container finished" podID="52516c81-59f5-4b50-9b16-80fcc8303cc9" containerID="70ff1201f7b65cc75fa50942282185e1f5b50ca891abbf21e514ecb3921235c6" exitCode=0 Dec 04 19:49:11 crc kubenswrapper[4733]: I1204 19:49:11.675423 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" event={"ID":"52516c81-59f5-4b50-9b16-80fcc8303cc9","Type":"ContainerDied","Data":"70ff1201f7b65cc75fa50942282185e1f5b50ca891abbf21e514ecb3921235c6"} Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.141782 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.190273 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ceph\") pod \"52516c81-59f5-4b50-9b16-80fcc8303cc9\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.190361 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ssh-key\") pod \"52516c81-59f5-4b50-9b16-80fcc8303cc9\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.190519 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-inventory\") pod \"52516c81-59f5-4b50-9b16-80fcc8303cc9\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.190547 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-676rn\" (UniqueName: \"kubernetes.io/projected/52516c81-59f5-4b50-9b16-80fcc8303cc9-kube-api-access-676rn\") pod \"52516c81-59f5-4b50-9b16-80fcc8303cc9\" (UID: \"52516c81-59f5-4b50-9b16-80fcc8303cc9\") " Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.196360 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52516c81-59f5-4b50-9b16-80fcc8303cc9-kube-api-access-676rn" (OuterVolumeSpecName: "kube-api-access-676rn") pod "52516c81-59f5-4b50-9b16-80fcc8303cc9" (UID: "52516c81-59f5-4b50-9b16-80fcc8303cc9"). InnerVolumeSpecName "kube-api-access-676rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.196405 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ceph" (OuterVolumeSpecName: "ceph") pod "52516c81-59f5-4b50-9b16-80fcc8303cc9" (UID: "52516c81-59f5-4b50-9b16-80fcc8303cc9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.219647 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "52516c81-59f5-4b50-9b16-80fcc8303cc9" (UID: "52516c81-59f5-4b50-9b16-80fcc8303cc9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.243317 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-inventory" (OuterVolumeSpecName: "inventory") pod "52516c81-59f5-4b50-9b16-80fcc8303cc9" (UID: "52516c81-59f5-4b50-9b16-80fcc8303cc9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.293442 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.293473 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.293485 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-676rn\" (UniqueName: \"kubernetes.io/projected/52516c81-59f5-4b50-9b16-80fcc8303cc9-kube-api-access-676rn\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.293493 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/52516c81-59f5-4b50-9b16-80fcc8303cc9-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.701718 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" event={"ID":"52516c81-59f5-4b50-9b16-80fcc8303cc9","Type":"ContainerDied","Data":"d27574df0ea0ed4c09fdaaa68d29d5944f43b9ba4ceb01bf750851bf55c0a450"} Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.701780 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d27574df0ea0ed4c09fdaaa68d29d5944f43b9ba4ceb01bf750851bf55c0a450" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.701903 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-ck6bp" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.780344 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-7w94z"] Dec 04 19:49:13 crc kubenswrapper[4733]: E1204 19:49:13.780999 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52516c81-59f5-4b50-9b16-80fcc8303cc9" containerName="reboot-os-openstack-openstack-cell1" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.781035 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="52516c81-59f5-4b50-9b16-80fcc8303cc9" containerName="reboot-os-openstack-openstack-cell1" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.781360 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="52516c81-59f5-4b50-9b16-80fcc8303cc9" containerName="reboot-os-openstack-openstack-cell1" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.782363 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.787460 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.787743 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.788649 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.789984 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.800695 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-7w94z"] Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.906973 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.907031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-inventory\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.907209 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ssh-key\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.907300 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.907335 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.907438 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.907716 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ceph\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.907834 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.907972 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.908138 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.908215 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:13 crc kubenswrapper[4733]: I1204 19:49:13.908391 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqvgm\" (UniqueName: \"kubernetes.io/projected/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-kube-api-access-tqvgm\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.010940 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011040 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011153 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqvgm\" (UniqueName: \"kubernetes.io/projected/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-kube-api-access-tqvgm\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011223 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011277 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-inventory\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011336 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ssh-key\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011387 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011423 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011502 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011615 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ceph\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011742 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.011875 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.018704 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ceph\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.018788 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ssh-key\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.019503 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.020010 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.023224 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.026127 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.029084 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.030751 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.031207 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.031459 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.033388 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-inventory\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.047023 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqvgm\" (UniqueName: \"kubernetes.io/projected/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-kube-api-access-tqvgm\") pod \"install-certs-openstack-openstack-cell1-7w94z\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:14 crc kubenswrapper[4733]: I1204 19:49:14.105060 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:15 crc kubenswrapper[4733]: I1204 19:49:14.699356 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-7w94z"] Dec 04 19:49:15 crc kubenswrapper[4733]: W1204 19:49:14.701517 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod271ce371_da5c_40a8_9f8f_bc90fa1bee8d.slice/crio-833e0cfab379f9cf95e2db539096b9361bca264801290616d928c547c43ed9c5 WatchSource:0}: Error finding container 833e0cfab379f9cf95e2db539096b9361bca264801290616d928c547c43ed9c5: Status 404 returned error can't find the container with id 833e0cfab379f9cf95e2db539096b9361bca264801290616d928c547c43ed9c5 Dec 04 19:49:15 crc kubenswrapper[4733]: I1204 19:49:14.716164 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" event={"ID":"271ce371-da5c-40a8-9f8f-bc90fa1bee8d","Type":"ContainerStarted","Data":"833e0cfab379f9cf95e2db539096b9361bca264801290616d928c547c43ed9c5"} Dec 04 19:49:15 crc kubenswrapper[4733]: I1204 19:49:15.361779 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:49:15 crc kubenswrapper[4733]: I1204 19:49:15.362172 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:49:15 crc kubenswrapper[4733]: I1204 19:49:15.724763 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" event={"ID":"271ce371-da5c-40a8-9f8f-bc90fa1bee8d","Type":"ContainerStarted","Data":"946f4f48623daf38697b1b7fd83318f89404e65a15176856b5c4d39fc70758f3"} Dec 04 19:49:15 crc kubenswrapper[4733]: I1204 19:49:15.759195 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" podStartSLOduration=2.258240972 podStartE2EDuration="2.759168884s" podCreationTimestamp="2025-12-04 19:49:13 +0000 UTC" firstStartedPulling="2025-12-04 19:49:14.704350953 +0000 UTC m=+7816.659711999" lastFinishedPulling="2025-12-04 19:49:15.205278875 +0000 UTC m=+7817.160639911" observedRunningTime="2025-12-04 19:49:15.750752436 +0000 UTC m=+7817.706113542" watchObservedRunningTime="2025-12-04 19:49:15.759168884 +0000 UTC m=+7817.714529930" Dec 04 19:49:34 crc kubenswrapper[4733]: I1204 19:49:34.924283 4733 generic.go:334] "Generic (PLEG): container finished" podID="271ce371-da5c-40a8-9f8f-bc90fa1bee8d" containerID="946f4f48623daf38697b1b7fd83318f89404e65a15176856b5c4d39fc70758f3" exitCode=0 Dec 04 19:49:34 crc kubenswrapper[4733]: I1204 19:49:34.924629 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" event={"ID":"271ce371-da5c-40a8-9f8f-bc90fa1bee8d","Type":"ContainerDied","Data":"946f4f48623daf38697b1b7fd83318f89404e65a15176856b5c4d39fc70758f3"} Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.433128 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.539869 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqvgm\" (UniqueName: \"kubernetes.io/projected/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-kube-api-access-tqvgm\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.539922 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-metadata-combined-ca-bundle\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.539942 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ovn-combined-ca-bundle\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.539979 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-dhcp-combined-ca-bundle\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.540005 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ssh-key\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.540090 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-inventory\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.540151 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-telemetry-combined-ca-bundle\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.540184 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-bootstrap-combined-ca-bundle\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.540270 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-nova-combined-ca-bundle\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.540346 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-sriov-combined-ca-bundle\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.540380 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-libvirt-combined-ca-bundle\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.540400 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ceph\") pod \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\" (UID: \"271ce371-da5c-40a8-9f8f-bc90fa1bee8d\") " Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.549187 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.549236 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.549265 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.549308 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.549471 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.549535 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.549571 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-kube-api-access-tqvgm" (OuterVolumeSpecName: "kube-api-access-tqvgm") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "kube-api-access-tqvgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.549817 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.551076 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.560024 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ceph" (OuterVolumeSpecName: "ceph") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.572362 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-inventory" (OuterVolumeSpecName: "inventory") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.597340 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "271ce371-da5c-40a8-9f8f-bc90fa1bee8d" (UID: "271ce371-da5c-40a8-9f8f-bc90fa1bee8d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642209 4733 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642243 4733 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642253 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642266 4733 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642276 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642287 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqvgm\" (UniqueName: \"kubernetes.io/projected/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-kube-api-access-tqvgm\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642297 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642307 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642316 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642324 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642332 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.642343 4733 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271ce371-da5c-40a8-9f8f-bc90fa1bee8d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.950069 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" event={"ID":"271ce371-da5c-40a8-9f8f-bc90fa1bee8d","Type":"ContainerDied","Data":"833e0cfab379f9cf95e2db539096b9361bca264801290616d928c547c43ed9c5"} Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.950134 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="833e0cfab379f9cf95e2db539096b9361bca264801290616d928c547c43ed9c5" Dec 04 19:49:36 crc kubenswrapper[4733]: I1204 19:49:36.950236 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-7w94z" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.060316 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-mpcsl"] Dec 04 19:49:37 crc kubenswrapper[4733]: E1204 19:49:37.060854 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271ce371-da5c-40a8-9f8f-bc90fa1bee8d" containerName="install-certs-openstack-openstack-cell1" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.060876 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="271ce371-da5c-40a8-9f8f-bc90fa1bee8d" containerName="install-certs-openstack-openstack-cell1" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.061116 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="271ce371-da5c-40a8-9f8f-bc90fa1bee8d" containerName="install-certs-openstack-openstack-cell1" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.061937 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.064395 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.064716 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.065004 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.065212 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.071290 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-mpcsl"] Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.153144 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.153292 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49l57\" (UniqueName: \"kubernetes.io/projected/2bbe511a-0c11-477f-a274-81556cc13c97-kube-api-access-49l57\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.153335 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-inventory\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.153627 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ceph\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.255395 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ceph\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.255515 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.255603 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49l57\" (UniqueName: \"kubernetes.io/projected/2bbe511a-0c11-477f-a274-81556cc13c97-kube-api-access-49l57\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.255641 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-inventory\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.260214 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-inventory\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.260407 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ceph\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.261231 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.277435 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49l57\" (UniqueName: \"kubernetes.io/projected/2bbe511a-0c11-477f-a274-81556cc13c97-kube-api-access-49l57\") pod \"ceph-client-openstack-openstack-cell1-mpcsl\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:37 crc kubenswrapper[4733]: I1204 19:49:37.387630 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:38 crc kubenswrapper[4733]: I1204 19:49:38.001182 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-mpcsl"] Dec 04 19:49:38 crc kubenswrapper[4733]: I1204 19:49:38.983277 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" event={"ID":"2bbe511a-0c11-477f-a274-81556cc13c97","Type":"ContainerStarted","Data":"c66ce841e0a27ffa920fcae6946bb98af33868474e0b31bc3b9eda7fce5d563a"} Dec 04 19:49:38 crc kubenswrapper[4733]: I1204 19:49:38.983998 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" event={"ID":"2bbe511a-0c11-477f-a274-81556cc13c97","Type":"ContainerStarted","Data":"df6d10bbaff6298ee89ed4f73ba209865b682c04edb49d12a275a87aab730af5"} Dec 04 19:49:39 crc kubenswrapper[4733]: I1204 19:49:39.026636 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" podStartSLOduration=1.5679643890000001 podStartE2EDuration="2.026609075s" podCreationTimestamp="2025-12-04 19:49:37 +0000 UTC" firstStartedPulling="2025-12-04 19:49:38.005294623 +0000 UTC m=+7839.960655689" lastFinishedPulling="2025-12-04 19:49:38.463939289 +0000 UTC m=+7840.419300375" observedRunningTime="2025-12-04 19:49:39.008055531 +0000 UTC m=+7840.963416637" watchObservedRunningTime="2025-12-04 19:49:39.026609075 +0000 UTC m=+7840.981970151" Dec 04 19:49:44 crc kubenswrapper[4733]: I1204 19:49:44.039640 4733 generic.go:334] "Generic (PLEG): container finished" podID="2bbe511a-0c11-477f-a274-81556cc13c97" containerID="c66ce841e0a27ffa920fcae6946bb98af33868474e0b31bc3b9eda7fce5d563a" exitCode=0 Dec 04 19:49:44 crc kubenswrapper[4733]: I1204 19:49:44.039892 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" event={"ID":"2bbe511a-0c11-477f-a274-81556cc13c97","Type":"ContainerDied","Data":"c66ce841e0a27ffa920fcae6946bb98af33868474e0b31bc3b9eda7fce5d563a"} Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.362589 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.363197 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.363289 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.365954 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.366122 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" gracePeriod=600 Dec 04 19:49:45 crc kubenswrapper[4733]: E1204 19:49:45.490897 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.587119 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.635360 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-inventory\") pod \"2bbe511a-0c11-477f-a274-81556cc13c97\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.635426 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49l57\" (UniqueName: \"kubernetes.io/projected/2bbe511a-0c11-477f-a274-81556cc13c97-kube-api-access-49l57\") pod \"2bbe511a-0c11-477f-a274-81556cc13c97\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.635475 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ceph\") pod \"2bbe511a-0c11-477f-a274-81556cc13c97\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.635512 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ssh-key\") pod \"2bbe511a-0c11-477f-a274-81556cc13c97\" (UID: \"2bbe511a-0c11-477f-a274-81556cc13c97\") " Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.642058 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbe511a-0c11-477f-a274-81556cc13c97-kube-api-access-49l57" (OuterVolumeSpecName: "kube-api-access-49l57") pod "2bbe511a-0c11-477f-a274-81556cc13c97" (UID: "2bbe511a-0c11-477f-a274-81556cc13c97"). InnerVolumeSpecName "kube-api-access-49l57". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.643605 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ceph" (OuterVolumeSpecName: "ceph") pod "2bbe511a-0c11-477f-a274-81556cc13c97" (UID: "2bbe511a-0c11-477f-a274-81556cc13c97"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.668950 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-inventory" (OuterVolumeSpecName: "inventory") pod "2bbe511a-0c11-477f-a274-81556cc13c97" (UID: "2bbe511a-0c11-477f-a274-81556cc13c97"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.669309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2bbe511a-0c11-477f-a274-81556cc13c97" (UID: "2bbe511a-0c11-477f-a274-81556cc13c97"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.739131 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.739358 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49l57\" (UniqueName: \"kubernetes.io/projected/2bbe511a-0c11-477f-a274-81556cc13c97-kube-api-access-49l57\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.739473 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:45 crc kubenswrapper[4733]: I1204 19:49:45.739545 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bbe511a-0c11-477f-a274-81556cc13c97-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.064991 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" exitCode=0 Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.065069 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d"} Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.065108 4733 scope.go:117] "RemoveContainer" containerID="6702c9ff250314030e3d19e8ac891392736fe3f613350f0fa8b792973e875133" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.065885 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:49:46 crc kubenswrapper[4733]: E1204 19:49:46.066784 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.067942 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" event={"ID":"2bbe511a-0c11-477f-a274-81556cc13c97","Type":"ContainerDied","Data":"df6d10bbaff6298ee89ed4f73ba209865b682c04edb49d12a275a87aab730af5"} Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.067964 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df6d10bbaff6298ee89ed4f73ba209865b682c04edb49d12a275a87aab730af5" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.068016 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-mpcsl" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.176804 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-ljcx8"] Dec 04 19:49:46 crc kubenswrapper[4733]: E1204 19:49:46.177845 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbe511a-0c11-477f-a274-81556cc13c97" containerName="ceph-client-openstack-openstack-cell1" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.177955 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbe511a-0c11-477f-a274-81556cc13c97" containerName="ceph-client-openstack-openstack-cell1" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.178341 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbe511a-0c11-477f-a274-81556cc13c97" containerName="ceph-client-openstack-openstack-cell1" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.179420 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.184006 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.184389 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.184585 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.184644 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.184731 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.196508 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-ljcx8"] Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.252177 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-inventory\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.252330 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrmhw\" (UniqueName: \"kubernetes.io/projected/a11a6634-6fde-4e35-9d88-a7def887dd5b-kube-api-access-zrmhw\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.252383 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ssh-key\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.252403 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.252437 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ceph\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.252715 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.355107 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrmhw\" (UniqueName: \"kubernetes.io/projected/a11a6634-6fde-4e35-9d88-a7def887dd5b-kube-api-access-zrmhw\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.355409 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ssh-key\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.355533 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.356388 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ceph\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.356550 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.356826 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-inventory\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.357305 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.360894 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ceph\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.362854 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.363192 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ssh-key\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.365867 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-inventory\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.377826 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrmhw\" (UniqueName: \"kubernetes.io/projected/a11a6634-6fde-4e35-9d88-a7def887dd5b-kube-api-access-zrmhw\") pod \"ovn-openstack-openstack-cell1-ljcx8\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:46 crc kubenswrapper[4733]: I1204 19:49:46.507369 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:49:47 crc kubenswrapper[4733]: I1204 19:49:47.070311 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-ljcx8"] Dec 04 19:49:47 crc kubenswrapper[4733]: I1204 19:49:47.090532 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" event={"ID":"a11a6634-6fde-4e35-9d88-a7def887dd5b","Type":"ContainerStarted","Data":"4b73d81774fb2447c11e1b7a78261672bdbd55241ed8c5029ed4d7b286eab060"} Dec 04 19:49:48 crc kubenswrapper[4733]: I1204 19:49:48.101735 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" event={"ID":"a11a6634-6fde-4e35-9d88-a7def887dd5b","Type":"ContainerStarted","Data":"f904ecabcc69577232e7930e3dd5a384697d0b1585407984cf07ff446a18dbcf"} Dec 04 19:49:48 crc kubenswrapper[4733]: I1204 19:49:48.129269 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" podStartSLOduration=1.666849105 podStartE2EDuration="2.129250683s" podCreationTimestamp="2025-12-04 19:49:46 +0000 UTC" firstStartedPulling="2025-12-04 19:49:47.066342873 +0000 UTC m=+7849.021703919" lastFinishedPulling="2025-12-04 19:49:47.528744441 +0000 UTC m=+7849.484105497" observedRunningTime="2025-12-04 19:49:48.127733541 +0000 UTC m=+7850.083094617" watchObservedRunningTime="2025-12-04 19:49:48.129250683 +0000 UTC m=+7850.084611749" Dec 04 19:49:57 crc kubenswrapper[4733]: I1204 19:49:57.337062 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:49:57 crc kubenswrapper[4733]: E1204 19:49:57.338261 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:50:10 crc kubenswrapper[4733]: I1204 19:50:10.336685 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:50:10 crc kubenswrapper[4733]: E1204 19:50:10.337780 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:50:23 crc kubenswrapper[4733]: I1204 19:50:23.337266 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:50:23 crc kubenswrapper[4733]: E1204 19:50:23.338405 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:50:36 crc kubenswrapper[4733]: I1204 19:50:36.335769 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:50:36 crc kubenswrapper[4733]: E1204 19:50:36.336466 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:50:47 crc kubenswrapper[4733]: I1204 19:50:47.337112 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:50:47 crc kubenswrapper[4733]: E1204 19:50:47.337869 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:50:58 crc kubenswrapper[4733]: I1204 19:50:58.892149 4733 generic.go:334] "Generic (PLEG): container finished" podID="a11a6634-6fde-4e35-9d88-a7def887dd5b" containerID="f904ecabcc69577232e7930e3dd5a384697d0b1585407984cf07ff446a18dbcf" exitCode=0 Dec 04 19:50:58 crc kubenswrapper[4733]: I1204 19:50:58.892274 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" event={"ID":"a11a6634-6fde-4e35-9d88-a7def887dd5b","Type":"ContainerDied","Data":"f904ecabcc69577232e7930e3dd5a384697d0b1585407984cf07ff446a18dbcf"} Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.340328 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:51:00 crc kubenswrapper[4733]: E1204 19:51:00.341247 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.354781 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.384968 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovncontroller-config-0\") pod \"a11a6634-6fde-4e35-9d88-a7def887dd5b\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.385143 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrmhw\" (UniqueName: \"kubernetes.io/projected/a11a6634-6fde-4e35-9d88-a7def887dd5b-kube-api-access-zrmhw\") pod \"a11a6634-6fde-4e35-9d88-a7def887dd5b\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.385378 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-inventory\") pod \"a11a6634-6fde-4e35-9d88-a7def887dd5b\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.385463 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ssh-key\") pod \"a11a6634-6fde-4e35-9d88-a7def887dd5b\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.385995 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovn-combined-ca-bundle\") pod \"a11a6634-6fde-4e35-9d88-a7def887dd5b\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.386061 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ceph\") pod \"a11a6634-6fde-4e35-9d88-a7def887dd5b\" (UID: \"a11a6634-6fde-4e35-9d88-a7def887dd5b\") " Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.400247 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a11a6634-6fde-4e35-9d88-a7def887dd5b-kube-api-access-zrmhw" (OuterVolumeSpecName: "kube-api-access-zrmhw") pod "a11a6634-6fde-4e35-9d88-a7def887dd5b" (UID: "a11a6634-6fde-4e35-9d88-a7def887dd5b"). InnerVolumeSpecName "kube-api-access-zrmhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.400477 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ceph" (OuterVolumeSpecName: "ceph") pod "a11a6634-6fde-4e35-9d88-a7def887dd5b" (UID: "a11a6634-6fde-4e35-9d88-a7def887dd5b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.403719 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a11a6634-6fde-4e35-9d88-a7def887dd5b" (UID: "a11a6634-6fde-4e35-9d88-a7def887dd5b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.420240 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-inventory" (OuterVolumeSpecName: "inventory") pod "a11a6634-6fde-4e35-9d88-a7def887dd5b" (UID: "a11a6634-6fde-4e35-9d88-a7def887dd5b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.421400 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a11a6634-6fde-4e35-9d88-a7def887dd5b" (UID: "a11a6634-6fde-4e35-9d88-a7def887dd5b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.435731 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "a11a6634-6fde-4e35-9d88-a7def887dd5b" (UID: "a11a6634-6fde-4e35-9d88-a7def887dd5b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.489266 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.489575 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.489589 4733 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a11a6634-6fde-4e35-9d88-a7def887dd5b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.489601 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrmhw\" (UniqueName: \"kubernetes.io/projected/a11a6634-6fde-4e35-9d88-a7def887dd5b-kube-api-access-zrmhw\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.489614 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.489625 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a11a6634-6fde-4e35-9d88-a7def887dd5b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.914571 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" event={"ID":"a11a6634-6fde-4e35-9d88-a7def887dd5b","Type":"ContainerDied","Data":"4b73d81774fb2447c11e1b7a78261672bdbd55241ed8c5029ed4d7b286eab060"} Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.914631 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b73d81774fb2447c11e1b7a78261672bdbd55241ed8c5029ed4d7b286eab060" Dec 04 19:51:00 crc kubenswrapper[4733]: I1204 19:51:00.914645 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-ljcx8" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.020073 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-kfjpj"] Dec 04 19:51:01 crc kubenswrapper[4733]: E1204 19:51:01.020534 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11a6634-6fde-4e35-9d88-a7def887dd5b" containerName="ovn-openstack-openstack-cell1" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.020551 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11a6634-6fde-4e35-9d88-a7def887dd5b" containerName="ovn-openstack-openstack-cell1" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.020782 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11a6634-6fde-4e35-9d88-a7def887dd5b" containerName="ovn-openstack-openstack-cell1" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.021574 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.026635 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.026658 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.027122 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.027347 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.027395 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.027907 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.035721 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-kfjpj"] Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.102494 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.102568 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.102668 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.102692 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.102830 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.103113 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.103178 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnpzd\" (UniqueName: \"kubernetes.io/projected/269e8fa2-f463-405c-8a43-8c083f6e6ea9-kube-api-access-wnpzd\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.205359 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.205510 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.205539 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnpzd\" (UniqueName: \"kubernetes.io/projected/269e8fa2-f463-405c-8a43-8c083f6e6ea9-kube-api-access-wnpzd\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.205568 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.205605 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.205670 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.205692 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.211573 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.211773 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.212236 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.212321 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.213559 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.224546 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.228602 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnpzd\" (UniqueName: \"kubernetes.io/projected/269e8fa2-f463-405c-8a43-8c083f6e6ea9-kube-api-access-wnpzd\") pod \"neutron-metadata-openstack-openstack-cell1-kfjpj\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.357379 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:51:01 crc kubenswrapper[4733]: I1204 19:51:01.987355 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-kfjpj"] Dec 04 19:51:02 crc kubenswrapper[4733]: I1204 19:51:02.936363 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" event={"ID":"269e8fa2-f463-405c-8a43-8c083f6e6ea9","Type":"ContainerStarted","Data":"31f272621fda1db84e7292b15f95fada378c891b3685d77b06ddfa9c8a8e0556"} Dec 04 19:51:03 crc kubenswrapper[4733]: I1204 19:51:03.946783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" event={"ID":"269e8fa2-f463-405c-8a43-8c083f6e6ea9","Type":"ContainerStarted","Data":"3288942d002cbf24f1e1055d1ba25e051fefa67978e0020069030b11a1724dd7"} Dec 04 19:51:03 crc kubenswrapper[4733]: I1204 19:51:03.964309 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" podStartSLOduration=3.182597657 podStartE2EDuration="3.964292592s" podCreationTimestamp="2025-12-04 19:51:00 +0000 UTC" firstStartedPulling="2025-12-04 19:51:01.982862907 +0000 UTC m=+7923.938223953" lastFinishedPulling="2025-12-04 19:51:02.764557842 +0000 UTC m=+7924.719918888" observedRunningTime="2025-12-04 19:51:03.961898757 +0000 UTC m=+7925.917259823" watchObservedRunningTime="2025-12-04 19:51:03.964292592 +0000 UTC m=+7925.919653638" Dec 04 19:51:13 crc kubenswrapper[4733]: I1204 19:51:13.338031 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:51:13 crc kubenswrapper[4733]: E1204 19:51:13.339135 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.805809 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v6znr"] Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.825149 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.846279 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmcj6\" (UniqueName: \"kubernetes.io/projected/1aed60fd-60cc-486a-9400-41bcc55e1228-kube-api-access-pmcj6\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.846333 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-catalog-content\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.846483 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-utilities\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.871863 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v6znr"] Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.949037 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-utilities\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.949151 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmcj6\" (UniqueName: \"kubernetes.io/projected/1aed60fd-60cc-486a-9400-41bcc55e1228-kube-api-access-pmcj6\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.949176 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-catalog-content\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.949648 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-catalog-content\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.949892 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-utilities\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:24 crc kubenswrapper[4733]: I1204 19:51:24.986655 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmcj6\" (UniqueName: \"kubernetes.io/projected/1aed60fd-60cc-486a-9400-41bcc55e1228-kube-api-access-pmcj6\") pod \"certified-operators-v6znr\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:25 crc kubenswrapper[4733]: I1204 19:51:25.185852 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:25 crc kubenswrapper[4733]: I1204 19:51:25.694837 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v6znr"] Dec 04 19:51:26 crc kubenswrapper[4733]: I1204 19:51:26.182970 4733 generic.go:334] "Generic (PLEG): container finished" podID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerID="7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2" exitCode=0 Dec 04 19:51:26 crc kubenswrapper[4733]: I1204 19:51:26.183377 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6znr" event={"ID":"1aed60fd-60cc-486a-9400-41bcc55e1228","Type":"ContainerDied","Data":"7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2"} Dec 04 19:51:26 crc kubenswrapper[4733]: I1204 19:51:26.183435 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6znr" event={"ID":"1aed60fd-60cc-486a-9400-41bcc55e1228","Type":"ContainerStarted","Data":"911bdc00488daaee10b6e4472fbeb5e807504552dc09b273d983da7750f17a9e"} Dec 04 19:51:27 crc kubenswrapper[4733]: I1204 19:51:27.336153 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:51:27 crc kubenswrapper[4733]: E1204 19:51:27.336679 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:51:28 crc kubenswrapper[4733]: I1204 19:51:28.206014 4733 generic.go:334] "Generic (PLEG): container finished" podID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerID="1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188" exitCode=0 Dec 04 19:51:28 crc kubenswrapper[4733]: I1204 19:51:28.206073 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6znr" event={"ID":"1aed60fd-60cc-486a-9400-41bcc55e1228","Type":"ContainerDied","Data":"1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188"} Dec 04 19:51:30 crc kubenswrapper[4733]: I1204 19:51:30.233590 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6znr" event={"ID":"1aed60fd-60cc-486a-9400-41bcc55e1228","Type":"ContainerStarted","Data":"30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68"} Dec 04 19:51:30 crc kubenswrapper[4733]: I1204 19:51:30.258077 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v6znr" podStartSLOduration=3.466536617 podStartE2EDuration="6.258058716s" podCreationTimestamp="2025-12-04 19:51:24 +0000 UTC" firstStartedPulling="2025-12-04 19:51:26.203638755 +0000 UTC m=+7948.158999801" lastFinishedPulling="2025-12-04 19:51:28.995160834 +0000 UTC m=+7950.950521900" observedRunningTime="2025-12-04 19:51:30.256042552 +0000 UTC m=+7952.211403618" watchObservedRunningTime="2025-12-04 19:51:30.258058716 +0000 UTC m=+7952.213419762" Dec 04 19:51:35 crc kubenswrapper[4733]: I1204 19:51:35.187035 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:35 crc kubenswrapper[4733]: I1204 19:51:35.188816 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:35 crc kubenswrapper[4733]: I1204 19:51:35.247986 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:35 crc kubenswrapper[4733]: I1204 19:51:35.374454 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:35 crc kubenswrapper[4733]: I1204 19:51:35.492527 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v6znr"] Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.334332 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v6znr" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerName="registry-server" containerID="cri-o://30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68" gracePeriod=2 Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.797149 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.858096 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmcj6\" (UniqueName: \"kubernetes.io/projected/1aed60fd-60cc-486a-9400-41bcc55e1228-kube-api-access-pmcj6\") pod \"1aed60fd-60cc-486a-9400-41bcc55e1228\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.858219 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-utilities\") pod \"1aed60fd-60cc-486a-9400-41bcc55e1228\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.858443 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-catalog-content\") pod \"1aed60fd-60cc-486a-9400-41bcc55e1228\" (UID: \"1aed60fd-60cc-486a-9400-41bcc55e1228\") " Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.859113 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-utilities" (OuterVolumeSpecName: "utilities") pod "1aed60fd-60cc-486a-9400-41bcc55e1228" (UID: "1aed60fd-60cc-486a-9400-41bcc55e1228"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.863697 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aed60fd-60cc-486a-9400-41bcc55e1228-kube-api-access-pmcj6" (OuterVolumeSpecName: "kube-api-access-pmcj6") pod "1aed60fd-60cc-486a-9400-41bcc55e1228" (UID: "1aed60fd-60cc-486a-9400-41bcc55e1228"). InnerVolumeSpecName "kube-api-access-pmcj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.917097 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1aed60fd-60cc-486a-9400-41bcc55e1228" (UID: "1aed60fd-60cc-486a-9400-41bcc55e1228"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.960422 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.960476 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmcj6\" (UniqueName: \"kubernetes.io/projected/1aed60fd-60cc-486a-9400-41bcc55e1228-kube-api-access-pmcj6\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:37 crc kubenswrapper[4733]: I1204 19:51:37.960491 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aed60fd-60cc-486a-9400-41bcc55e1228-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.356749 4733 generic.go:334] "Generic (PLEG): container finished" podID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerID="30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68" exitCode=0 Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.356983 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6znr" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.357004 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6znr" event={"ID":"1aed60fd-60cc-486a-9400-41bcc55e1228","Type":"ContainerDied","Data":"30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68"} Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.358258 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6znr" event={"ID":"1aed60fd-60cc-486a-9400-41bcc55e1228","Type":"ContainerDied","Data":"911bdc00488daaee10b6e4472fbeb5e807504552dc09b273d983da7750f17a9e"} Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.358355 4733 scope.go:117] "RemoveContainer" containerID="30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.396470 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v6znr"] Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.401313 4733 scope.go:117] "RemoveContainer" containerID="1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.413384 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v6znr"] Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.425442 4733 scope.go:117] "RemoveContainer" containerID="7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.482768 4733 scope.go:117] "RemoveContainer" containerID="30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68" Dec 04 19:51:38 crc kubenswrapper[4733]: E1204 19:51:38.483343 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68\": container with ID starting with 30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68 not found: ID does not exist" containerID="30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.483399 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68"} err="failed to get container status \"30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68\": rpc error: code = NotFound desc = could not find container \"30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68\": container with ID starting with 30901463f37c6613eaaebe8c274b3ddbbb3bf2718a75154dacac3270a03bec68 not found: ID does not exist" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.483432 4733 scope.go:117] "RemoveContainer" containerID="1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188" Dec 04 19:51:38 crc kubenswrapper[4733]: E1204 19:51:38.484194 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188\": container with ID starting with 1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188 not found: ID does not exist" containerID="1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.484227 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188"} err="failed to get container status \"1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188\": rpc error: code = NotFound desc = could not find container \"1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188\": container with ID starting with 1ec3aadaf61e420926a72ac31d53dc9c9a3307310fd118341f4b7681d93d8188 not found: ID does not exist" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.484294 4733 scope.go:117] "RemoveContainer" containerID="7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2" Dec 04 19:51:38 crc kubenswrapper[4733]: E1204 19:51:38.484590 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2\": container with ID starting with 7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2 not found: ID does not exist" containerID="7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2" Dec 04 19:51:38 crc kubenswrapper[4733]: I1204 19:51:38.484623 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2"} err="failed to get container status \"7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2\": rpc error: code = NotFound desc = could not find container \"7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2\": container with ID starting with 7ca097cdb39e7c97ec3ca400131ca9d182d95bd7a24d9561c7e7b467787b19c2 not found: ID does not exist" Dec 04 19:51:40 crc kubenswrapper[4733]: I1204 19:51:40.335728 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:51:40 crc kubenswrapper[4733]: E1204 19:51:40.336313 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:51:40 crc kubenswrapper[4733]: I1204 19:51:40.358213 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" path="/var/lib/kubelet/pods/1aed60fd-60cc-486a-9400-41bcc55e1228/volumes" Dec 04 19:51:54 crc kubenswrapper[4733]: I1204 19:51:54.338593 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:51:54 crc kubenswrapper[4733]: E1204 19:51:54.339530 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:51:59 crc kubenswrapper[4733]: I1204 19:51:59.582765 4733 generic.go:334] "Generic (PLEG): container finished" podID="269e8fa2-f463-405c-8a43-8c083f6e6ea9" containerID="3288942d002cbf24f1e1055d1ba25e051fefa67978e0020069030b11a1724dd7" exitCode=0 Dec 04 19:51:59 crc kubenswrapper[4733]: I1204 19:51:59.582834 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" event={"ID":"269e8fa2-f463-405c-8a43-8c083f6e6ea9","Type":"ContainerDied","Data":"3288942d002cbf24f1e1055d1ba25e051fefa67978e0020069030b11a1724dd7"} Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.068155 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.204977 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.205276 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-inventory\") pod \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.205431 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-nova-metadata-neutron-config-0\") pod \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.205625 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ceph\") pod \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.206050 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-metadata-combined-ca-bundle\") pod \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.206317 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnpzd\" (UniqueName: \"kubernetes.io/projected/269e8fa2-f463-405c-8a43-8c083f6e6ea9-kube-api-access-wnpzd\") pod \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.206527 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ssh-key\") pod \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\" (UID: \"269e8fa2-f463-405c-8a43-8c083f6e6ea9\") " Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.213748 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ceph" (OuterVolumeSpecName: "ceph") pod "269e8fa2-f463-405c-8a43-8c083f6e6ea9" (UID: "269e8fa2-f463-405c-8a43-8c083f6e6ea9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.215225 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "269e8fa2-f463-405c-8a43-8c083f6e6ea9" (UID: "269e8fa2-f463-405c-8a43-8c083f6e6ea9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.218071 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/269e8fa2-f463-405c-8a43-8c083f6e6ea9-kube-api-access-wnpzd" (OuterVolumeSpecName: "kube-api-access-wnpzd") pod "269e8fa2-f463-405c-8a43-8c083f6e6ea9" (UID: "269e8fa2-f463-405c-8a43-8c083f6e6ea9"). InnerVolumeSpecName "kube-api-access-wnpzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.253322 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "269e8fa2-f463-405c-8a43-8c083f6e6ea9" (UID: "269e8fa2-f463-405c-8a43-8c083f6e6ea9"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.254070 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-inventory" (OuterVolumeSpecName: "inventory") pod "269e8fa2-f463-405c-8a43-8c083f6e6ea9" (UID: "269e8fa2-f463-405c-8a43-8c083f6e6ea9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.254781 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "269e8fa2-f463-405c-8a43-8c083f6e6ea9" (UID: "269e8fa2-f463-405c-8a43-8c083f6e6ea9"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.264313 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "269e8fa2-f463-405c-8a43-8c083f6e6ea9" (UID: "269e8fa2-f463-405c-8a43-8c083f6e6ea9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.310942 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.310978 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnpzd\" (UniqueName: \"kubernetes.io/projected/269e8fa2-f463-405c-8a43-8c083f6e6ea9-kube-api-access-wnpzd\") on node \"crc\" DevicePath \"\"" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.310987 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.310997 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.311008 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.311017 4733 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.311026 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/269e8fa2-f463-405c-8a43-8c083f6e6ea9-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.605943 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" event={"ID":"269e8fa2-f463-405c-8a43-8c083f6e6ea9","Type":"ContainerDied","Data":"31f272621fda1db84e7292b15f95fada378c891b3685d77b06ddfa9c8a8e0556"} Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.606319 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31f272621fda1db84e7292b15f95fada378c891b3685d77b06ddfa9c8a8e0556" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.606242 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-kfjpj" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.708068 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zk2ps"] Dec 04 19:52:01 crc kubenswrapper[4733]: E1204 19:52:01.708501 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerName="extract-utilities" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.708519 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerName="extract-utilities" Dec 04 19:52:01 crc kubenswrapper[4733]: E1204 19:52:01.708547 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerName="registry-server" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.708554 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerName="registry-server" Dec 04 19:52:01 crc kubenswrapper[4733]: E1204 19:52:01.708564 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerName="extract-content" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.708570 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerName="extract-content" Dec 04 19:52:01 crc kubenswrapper[4733]: E1204 19:52:01.708587 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269e8fa2-f463-405c-8a43-8c083f6e6ea9" containerName="neutron-metadata-openstack-openstack-cell1" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.708594 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="269e8fa2-f463-405c-8a43-8c083f6e6ea9" containerName="neutron-metadata-openstack-openstack-cell1" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.708819 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="269e8fa2-f463-405c-8a43-8c083f6e6ea9" containerName="neutron-metadata-openstack-openstack-cell1" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.708839 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aed60fd-60cc-486a-9400-41bcc55e1228" containerName="registry-server" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.709560 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.711449 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.711782 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.712166 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.712380 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.713207 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.726086 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zk2ps"] Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.821438 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-inventory\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.821520 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmh8z\" (UniqueName: \"kubernetes.io/projected/df93d53f-da19-405c-a337-bbce7af8fd33-kube-api-access-xmh8z\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.821571 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.821819 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ceph\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.822180 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.822246 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.924684 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.924754 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.924842 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-inventory\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.925024 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmh8z\" (UniqueName: \"kubernetes.io/projected/df93d53f-da19-405c-a337-bbce7af8fd33-kube-api-access-xmh8z\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.925139 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.925258 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ceph\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.928651 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.928729 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ceph\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.928999 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-inventory\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.929449 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.929507 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:01 crc kubenswrapper[4733]: I1204 19:52:01.945850 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmh8z\" (UniqueName: \"kubernetes.io/projected/df93d53f-da19-405c-a337-bbce7af8fd33-kube-api-access-xmh8z\") pod \"libvirt-openstack-openstack-cell1-zk2ps\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:02 crc kubenswrapper[4733]: I1204 19:52:02.028619 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:52:02 crc kubenswrapper[4733]: I1204 19:52:02.637611 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zk2ps"] Dec 04 19:52:02 crc kubenswrapper[4733]: W1204 19:52:02.646286 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf93d53f_da19_405c_a337_bbce7af8fd33.slice/crio-cfe295ea748a0690aa6647bb64cb1f67958dc002cbe7e8a92105cb5f6962623c WatchSource:0}: Error finding container cfe295ea748a0690aa6647bb64cb1f67958dc002cbe7e8a92105cb5f6962623c: Status 404 returned error can't find the container with id cfe295ea748a0690aa6647bb64cb1f67958dc002cbe7e8a92105cb5f6962623c Dec 04 19:52:02 crc kubenswrapper[4733]: I1204 19:52:02.649433 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 19:52:03 crc kubenswrapper[4733]: I1204 19:52:03.630065 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" event={"ID":"df93d53f-da19-405c-a337-bbce7af8fd33","Type":"ContainerStarted","Data":"3730731e9093a3c0f028e1653e5a916fd37ed0f4cba3f31192dce7b4f57c6e7a"} Dec 04 19:52:03 crc kubenswrapper[4733]: I1204 19:52:03.630405 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" event={"ID":"df93d53f-da19-405c-a337-bbce7af8fd33","Type":"ContainerStarted","Data":"cfe295ea748a0690aa6647bb64cb1f67958dc002cbe7e8a92105cb5f6962623c"} Dec 04 19:52:03 crc kubenswrapper[4733]: I1204 19:52:03.662141 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" podStartSLOduration=2.173771629 podStartE2EDuration="2.66212507s" podCreationTimestamp="2025-12-04 19:52:01 +0000 UTC" firstStartedPulling="2025-12-04 19:52:02.649132804 +0000 UTC m=+7984.604493860" lastFinishedPulling="2025-12-04 19:52:03.137486255 +0000 UTC m=+7985.092847301" observedRunningTime="2025-12-04 19:52:03.658850011 +0000 UTC m=+7985.614211077" watchObservedRunningTime="2025-12-04 19:52:03.66212507 +0000 UTC m=+7985.617486116" Dec 04 19:52:06 crc kubenswrapper[4733]: I1204 19:52:06.335726 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:52:06 crc kubenswrapper[4733]: E1204 19:52:06.336458 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:52:18 crc kubenswrapper[4733]: I1204 19:52:18.355495 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:52:18 crc kubenswrapper[4733]: E1204 19:52:18.356468 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:52:32 crc kubenswrapper[4733]: I1204 19:52:32.337295 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:52:32 crc kubenswrapper[4733]: E1204 19:52:32.338183 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:52:45 crc kubenswrapper[4733]: I1204 19:52:45.335749 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:52:45 crc kubenswrapper[4733]: E1204 19:52:45.336741 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:52:57 crc kubenswrapper[4733]: I1204 19:52:57.335525 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:52:57 crc kubenswrapper[4733]: E1204 19:52:57.336605 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:53:12 crc kubenswrapper[4733]: I1204 19:53:12.335634 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:53:12 crc kubenswrapper[4733]: E1204 19:53:12.336678 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:53:25 crc kubenswrapper[4733]: I1204 19:53:25.336366 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:53:25 crc kubenswrapper[4733]: E1204 19:53:25.337720 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:53:37 crc kubenswrapper[4733]: I1204 19:53:37.336122 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:53:37 crc kubenswrapper[4733]: E1204 19:53:37.337893 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:53:51 crc kubenswrapper[4733]: I1204 19:53:51.335590 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:53:51 crc kubenswrapper[4733]: E1204 19:53:51.337121 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:54:06 crc kubenswrapper[4733]: I1204 19:54:06.336066 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:54:06 crc kubenswrapper[4733]: E1204 19:54:06.337293 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:54:20 crc kubenswrapper[4733]: I1204 19:54:20.336567 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:54:20 crc kubenswrapper[4733]: E1204 19:54:20.337517 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:54:32 crc kubenswrapper[4733]: I1204 19:54:32.335736 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:54:32 crc kubenswrapper[4733]: E1204 19:54:32.336547 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:54:45 crc kubenswrapper[4733]: I1204 19:54:45.336016 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:54:45 crc kubenswrapper[4733]: E1204 19:54:45.337358 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 19:54:56 crc kubenswrapper[4733]: I1204 19:54:56.336678 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:54:57 crc kubenswrapper[4733]: I1204 19:54:57.176154 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"06b77c5359e2bbbe0c6023f6861cdc478932b407b029be4bc868ca52a499a232"} Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.730579 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z6qb2"] Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.734033 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.755171 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z6qb2"] Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.870443 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-catalog-content\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.870520 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgtb9\" (UniqueName: \"kubernetes.io/projected/f5dd5103-4702-4d44-8f45-95babdafc86d-kube-api-access-mgtb9\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.870569 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-utilities\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.972456 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-catalog-content\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.972556 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgtb9\" (UniqueName: \"kubernetes.io/projected/f5dd5103-4702-4d44-8f45-95babdafc86d-kube-api-access-mgtb9\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.972606 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-utilities\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.973364 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-utilities\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.973364 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-catalog-content\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:39 crc kubenswrapper[4733]: I1204 19:55:39.999832 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgtb9\" (UniqueName: \"kubernetes.io/projected/f5dd5103-4702-4d44-8f45-95babdafc86d-kube-api-access-mgtb9\") pod \"redhat-marketplace-z6qb2\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:40 crc kubenswrapper[4733]: I1204 19:55:40.070582 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:40 crc kubenswrapper[4733]: I1204 19:55:40.619023 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z6qb2"] Dec 04 19:55:40 crc kubenswrapper[4733]: I1204 19:55:40.751751 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z6qb2" event={"ID":"f5dd5103-4702-4d44-8f45-95babdafc86d","Type":"ContainerStarted","Data":"26f5bd34ec46ae1354b39bad67c7a0b43dde59ad31f140f05156e7c208697f02"} Dec 04 19:55:41 crc kubenswrapper[4733]: I1204 19:55:41.764233 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerID="be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea" exitCode=0 Dec 04 19:55:41 crc kubenswrapper[4733]: I1204 19:55:41.764391 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z6qb2" event={"ID":"f5dd5103-4702-4d44-8f45-95babdafc86d","Type":"ContainerDied","Data":"be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea"} Dec 04 19:55:42 crc kubenswrapper[4733]: I1204 19:55:42.775117 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z6qb2" event={"ID":"f5dd5103-4702-4d44-8f45-95babdafc86d","Type":"ContainerStarted","Data":"cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52"} Dec 04 19:55:43 crc kubenswrapper[4733]: I1204 19:55:43.791212 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerID="cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52" exitCode=0 Dec 04 19:55:43 crc kubenswrapper[4733]: I1204 19:55:43.791478 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z6qb2" event={"ID":"f5dd5103-4702-4d44-8f45-95babdafc86d","Type":"ContainerDied","Data":"cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52"} Dec 04 19:55:44 crc kubenswrapper[4733]: I1204 19:55:44.810780 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z6qb2" event={"ID":"f5dd5103-4702-4d44-8f45-95babdafc86d","Type":"ContainerStarted","Data":"45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe"} Dec 04 19:55:44 crc kubenswrapper[4733]: I1204 19:55:44.849606 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z6qb2" podStartSLOduration=3.398719925 podStartE2EDuration="5.849582107s" podCreationTimestamp="2025-12-04 19:55:39 +0000 UTC" firstStartedPulling="2025-12-04 19:55:41.767081779 +0000 UTC m=+8203.722442825" lastFinishedPulling="2025-12-04 19:55:44.217943941 +0000 UTC m=+8206.173305007" observedRunningTime="2025-12-04 19:55:44.835769072 +0000 UTC m=+8206.791130158" watchObservedRunningTime="2025-12-04 19:55:44.849582107 +0000 UTC m=+8206.804943163" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.581929 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m499p"] Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.584409 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.587551 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-utilities\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.587589 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-catalog-content\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.587845 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfm8r\" (UniqueName: \"kubernetes.io/projected/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-kube-api-access-mfm8r\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.599843 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m499p"] Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.689940 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfm8r\" (UniqueName: \"kubernetes.io/projected/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-kube-api-access-mfm8r\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.690093 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-utilities\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.690120 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-catalog-content\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.690564 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-catalog-content\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.690632 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-utilities\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.711552 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfm8r\" (UniqueName: \"kubernetes.io/projected/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-kube-api-access-mfm8r\") pod \"community-operators-m499p\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:48 crc kubenswrapper[4733]: I1204 19:55:48.921300 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:49 crc kubenswrapper[4733]: I1204 19:55:49.590090 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m499p"] Dec 04 19:55:49 crc kubenswrapper[4733]: I1204 19:55:49.892856 4733 generic.go:334] "Generic (PLEG): container finished" podID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerID="2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee" exitCode=0 Dec 04 19:55:49 crc kubenswrapper[4733]: I1204 19:55:49.892896 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m499p" event={"ID":"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d","Type":"ContainerDied","Data":"2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee"} Dec 04 19:55:49 crc kubenswrapper[4733]: I1204 19:55:49.892922 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m499p" event={"ID":"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d","Type":"ContainerStarted","Data":"5eb74d4d0185c0e00c445be90970240c30303000a920c96044117045b626e38c"} Dec 04 19:55:50 crc kubenswrapper[4733]: I1204 19:55:50.070775 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:50 crc kubenswrapper[4733]: I1204 19:55:50.071090 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:50 crc kubenswrapper[4733]: I1204 19:55:50.139090 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:50 crc kubenswrapper[4733]: I1204 19:55:50.976917 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:51 crc kubenswrapper[4733]: I1204 19:55:51.915882 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m499p" event={"ID":"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d","Type":"ContainerStarted","Data":"1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199"} Dec 04 19:55:52 crc kubenswrapper[4733]: I1204 19:55:52.550696 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z6qb2"] Dec 04 19:55:52 crc kubenswrapper[4733]: I1204 19:55:52.935357 4733 generic.go:334] "Generic (PLEG): container finished" podID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerID="1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199" exitCode=0 Dec 04 19:55:52 crc kubenswrapper[4733]: I1204 19:55:52.935550 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z6qb2" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerName="registry-server" containerID="cri-o://45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe" gracePeriod=2 Dec 04 19:55:52 crc kubenswrapper[4733]: I1204 19:55:52.936603 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m499p" event={"ID":"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d","Type":"ContainerDied","Data":"1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199"} Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.559734 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.608306 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-catalog-content\") pod \"f5dd5103-4702-4d44-8f45-95babdafc86d\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.608453 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgtb9\" (UniqueName: \"kubernetes.io/projected/f5dd5103-4702-4d44-8f45-95babdafc86d-kube-api-access-mgtb9\") pod \"f5dd5103-4702-4d44-8f45-95babdafc86d\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.608505 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-utilities\") pod \"f5dd5103-4702-4d44-8f45-95babdafc86d\" (UID: \"f5dd5103-4702-4d44-8f45-95babdafc86d\") " Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.609540 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-utilities" (OuterVolumeSpecName: "utilities") pod "f5dd5103-4702-4d44-8f45-95babdafc86d" (UID: "f5dd5103-4702-4d44-8f45-95babdafc86d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.626224 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5dd5103-4702-4d44-8f45-95babdafc86d" (UID: "f5dd5103-4702-4d44-8f45-95babdafc86d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.629547 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5dd5103-4702-4d44-8f45-95babdafc86d-kube-api-access-mgtb9" (OuterVolumeSpecName: "kube-api-access-mgtb9") pod "f5dd5103-4702-4d44-8f45-95babdafc86d" (UID: "f5dd5103-4702-4d44-8f45-95babdafc86d"). InnerVolumeSpecName "kube-api-access-mgtb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.711449 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.711491 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgtb9\" (UniqueName: \"kubernetes.io/projected/f5dd5103-4702-4d44-8f45-95babdafc86d-kube-api-access-mgtb9\") on node \"crc\" DevicePath \"\"" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.711507 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dd5103-4702-4d44-8f45-95babdafc86d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.949443 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerID="45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe" exitCode=0 Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.949508 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z6qb2" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.949522 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z6qb2" event={"ID":"f5dd5103-4702-4d44-8f45-95babdafc86d","Type":"ContainerDied","Data":"45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe"} Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.949564 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z6qb2" event={"ID":"f5dd5103-4702-4d44-8f45-95babdafc86d","Type":"ContainerDied","Data":"26f5bd34ec46ae1354b39bad67c7a0b43dde59ad31f140f05156e7c208697f02"} Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.949581 4733 scope.go:117] "RemoveContainer" containerID="45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.954722 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m499p" event={"ID":"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d","Type":"ContainerStarted","Data":"d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3"} Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.974141 4733 scope.go:117] "RemoveContainer" containerID="cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52" Dec 04 19:55:53 crc kubenswrapper[4733]: I1204 19:55:53.981549 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m499p" podStartSLOduration=2.498486422 podStartE2EDuration="5.981535251s" podCreationTimestamp="2025-12-04 19:55:48 +0000 UTC" firstStartedPulling="2025-12-04 19:55:49.894466774 +0000 UTC m=+8211.849827830" lastFinishedPulling="2025-12-04 19:55:53.377515603 +0000 UTC m=+8215.332876659" observedRunningTime="2025-12-04 19:55:53.979923997 +0000 UTC m=+8215.935285053" watchObservedRunningTime="2025-12-04 19:55:53.981535251 +0000 UTC m=+8215.936896297" Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.010159 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z6qb2"] Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.020430 4733 scope.go:117] "RemoveContainer" containerID="be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea" Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.048203 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z6qb2"] Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.079104 4733 scope.go:117] "RemoveContainer" containerID="45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe" Dec 04 19:55:54 crc kubenswrapper[4733]: E1204 19:55:54.082056 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe\": container with ID starting with 45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe not found: ID does not exist" containerID="45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe" Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.082129 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe"} err="failed to get container status \"45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe\": rpc error: code = NotFound desc = could not find container \"45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe\": container with ID starting with 45f80a4a1128694e91c45bddbef1d58348e5e5179b1f2f0761d826bbaf361fbe not found: ID does not exist" Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.082165 4733 scope.go:117] "RemoveContainer" containerID="cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52" Dec 04 19:55:54 crc kubenswrapper[4733]: E1204 19:55:54.082638 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52\": container with ID starting with cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52 not found: ID does not exist" containerID="cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52" Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.082680 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52"} err="failed to get container status \"cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52\": rpc error: code = NotFound desc = could not find container \"cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52\": container with ID starting with cdf0955d709882428b1257a0b2816ccc5b91115243a71e96d7f3dc9533be1b52 not found: ID does not exist" Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.082710 4733 scope.go:117] "RemoveContainer" containerID="be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea" Dec 04 19:55:54 crc kubenswrapper[4733]: E1204 19:55:54.083022 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea\": container with ID starting with be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea not found: ID does not exist" containerID="be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea" Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.083054 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea"} err="failed to get container status \"be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea\": rpc error: code = NotFound desc = could not find container \"be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea\": container with ID starting with be8042802c4c914c0cdfab447e85370ce962ffd3253cc6018b6abdfc02bd1cea not found: ID does not exist" Dec 04 19:55:54 crc kubenswrapper[4733]: I1204 19:55:54.351595 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" path="/var/lib/kubelet/pods/f5dd5103-4702-4d44-8f45-95babdafc86d/volumes" Dec 04 19:55:58 crc kubenswrapper[4733]: I1204 19:55:58.923377 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:58 crc kubenswrapper[4733]: I1204 19:55:58.923786 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:58 crc kubenswrapper[4733]: I1204 19:55:58.995419 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:59 crc kubenswrapper[4733]: I1204 19:55:59.089303 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m499p" Dec 04 19:55:59 crc kubenswrapper[4733]: I1204 19:55:59.244013 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m499p"] Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.052004 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m499p" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerName="registry-server" containerID="cri-o://d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3" gracePeriod=2 Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.724519 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m499p" Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.845486 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfm8r\" (UniqueName: \"kubernetes.io/projected/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-kube-api-access-mfm8r\") pod \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.845538 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-catalog-content\") pod \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.845660 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-utilities\") pod \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\" (UID: \"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d\") " Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.846307 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-utilities" (OuterVolumeSpecName: "utilities") pod "ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" (UID: "ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.852178 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-kube-api-access-mfm8r" (OuterVolumeSpecName: "kube-api-access-mfm8r") pod "ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" (UID: "ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d"). InnerVolumeSpecName "kube-api-access-mfm8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.903903 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" (UID: "ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.947403 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfm8r\" (UniqueName: \"kubernetes.io/projected/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-kube-api-access-mfm8r\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.947585 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:01 crc kubenswrapper[4733]: I1204 19:56:01.947693 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.071837 4733 generic.go:334] "Generic (PLEG): container finished" podID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerID="d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3" exitCode=0 Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.071911 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m499p" event={"ID":"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d","Type":"ContainerDied","Data":"d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3"} Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.071963 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m499p" event={"ID":"ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d","Type":"ContainerDied","Data":"5eb74d4d0185c0e00c445be90970240c30303000a920c96044117045b626e38c"} Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.071979 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m499p" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.072001 4733 scope.go:117] "RemoveContainer" containerID="d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.098044 4733 scope.go:117] "RemoveContainer" containerID="1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.143341 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m499p"] Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.158659 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m499p"] Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.159438 4733 scope.go:117] "RemoveContainer" containerID="2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.196937 4733 scope.go:117] "RemoveContainer" containerID="d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3" Dec 04 19:56:02 crc kubenswrapper[4733]: E1204 19:56:02.197501 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3\": container with ID starting with d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3 not found: ID does not exist" containerID="d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.197554 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3"} err="failed to get container status \"d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3\": rpc error: code = NotFound desc = could not find container \"d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3\": container with ID starting with d954ddc1d11c30901baee0af8a1d8fd5de224f169b33574239dfc9960bb4c8d3 not found: ID does not exist" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.197588 4733 scope.go:117] "RemoveContainer" containerID="1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199" Dec 04 19:56:02 crc kubenswrapper[4733]: E1204 19:56:02.197937 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199\": container with ID starting with 1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199 not found: ID does not exist" containerID="1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.197977 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199"} err="failed to get container status \"1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199\": rpc error: code = NotFound desc = could not find container \"1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199\": container with ID starting with 1ae81a1f673768f6af454b3dfec7ecbf6d209c1a7d587eed1cba90e557381199 not found: ID does not exist" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.198003 4733 scope.go:117] "RemoveContainer" containerID="2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee" Dec 04 19:56:02 crc kubenswrapper[4733]: E1204 19:56:02.209581 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee\": container with ID starting with 2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee not found: ID does not exist" containerID="2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.209618 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee"} err="failed to get container status \"2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee\": rpc error: code = NotFound desc = could not find container \"2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee\": container with ID starting with 2b43569df17cc923999ca1eb591efc6bd4dc840dd1ba7d9816c51aff18feb4ee not found: ID does not exist" Dec 04 19:56:02 crc kubenswrapper[4733]: I1204 19:56:02.363566 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" path="/var/lib/kubelet/pods/ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d/volumes" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.935458 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d4qbm"] Dec 04 19:56:18 crc kubenswrapper[4733]: E1204 19:56:18.938959 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerName="extract-utilities" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.939075 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerName="extract-utilities" Dec 04 19:56:18 crc kubenswrapper[4733]: E1204 19:56:18.939155 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerName="extract-content" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.939226 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerName="extract-content" Dec 04 19:56:18 crc kubenswrapper[4733]: E1204 19:56:18.939331 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerName="extract-utilities" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.939407 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerName="extract-utilities" Dec 04 19:56:18 crc kubenswrapper[4733]: E1204 19:56:18.939489 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerName="registry-server" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.939558 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerName="registry-server" Dec 04 19:56:18 crc kubenswrapper[4733]: E1204 19:56:18.939640 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerName="registry-server" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.939711 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerName="registry-server" Dec 04 19:56:18 crc kubenswrapper[4733]: E1204 19:56:18.939826 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerName="extract-content" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.939905 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerName="extract-content" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.940465 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5dd5103-4702-4d44-8f45-95babdafc86d" containerName="registry-server" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.940568 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad091e75-cb3d-4d7e-b7e7-dd4e77aaf75d" containerName="registry-server" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.943755 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:18 crc kubenswrapper[4733]: I1204 19:56:18.962235 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4qbm"] Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.082886 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-utilities\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.083202 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-645gw\" (UniqueName: \"kubernetes.io/projected/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-kube-api-access-645gw\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.083299 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-catalog-content\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.185531 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-645gw\" (UniqueName: \"kubernetes.io/projected/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-kube-api-access-645gw\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.185672 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-catalog-content\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.185764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-utilities\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.186417 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-utilities\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.186434 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-catalog-content\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.217489 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-645gw\" (UniqueName: \"kubernetes.io/projected/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-kube-api-access-645gw\") pod \"redhat-operators-d4qbm\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.273213 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:19 crc kubenswrapper[4733]: I1204 19:56:19.861066 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4qbm"] Dec 04 19:56:20 crc kubenswrapper[4733]: I1204 19:56:20.282276 4733 generic.go:334] "Generic (PLEG): container finished" podID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerID="0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da" exitCode=0 Dec 04 19:56:20 crc kubenswrapper[4733]: I1204 19:56:20.282381 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4qbm" event={"ID":"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde","Type":"ContainerDied","Data":"0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da"} Dec 04 19:56:20 crc kubenswrapper[4733]: I1204 19:56:20.282632 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4qbm" event={"ID":"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde","Type":"ContainerStarted","Data":"ca382d07398b9b6df253005dcb35286c126caa1d32dfb7e0d39b449405514ffa"} Dec 04 19:56:22 crc kubenswrapper[4733]: I1204 19:56:22.309289 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4qbm" event={"ID":"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde","Type":"ContainerStarted","Data":"e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9"} Dec 04 19:56:25 crc kubenswrapper[4733]: I1204 19:56:25.344194 4733 generic.go:334] "Generic (PLEG): container finished" podID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerID="e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9" exitCode=0 Dec 04 19:56:25 crc kubenswrapper[4733]: I1204 19:56:25.344311 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4qbm" event={"ID":"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde","Type":"ContainerDied","Data":"e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9"} Dec 04 19:56:26 crc kubenswrapper[4733]: I1204 19:56:26.357971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4qbm" event={"ID":"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde","Type":"ContainerStarted","Data":"b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7"} Dec 04 19:56:26 crc kubenswrapper[4733]: I1204 19:56:26.387055 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d4qbm" podStartSLOduration=2.696632692 podStartE2EDuration="8.387031311s" podCreationTimestamp="2025-12-04 19:56:18 +0000 UTC" firstStartedPulling="2025-12-04 19:56:20.284499086 +0000 UTC m=+8242.239860132" lastFinishedPulling="2025-12-04 19:56:25.974897705 +0000 UTC m=+8247.930258751" observedRunningTime="2025-12-04 19:56:26.375648372 +0000 UTC m=+8248.331009438" watchObservedRunningTime="2025-12-04 19:56:26.387031311 +0000 UTC m=+8248.342392367" Dec 04 19:56:29 crc kubenswrapper[4733]: I1204 19:56:29.274100 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:29 crc kubenswrapper[4733]: I1204 19:56:29.274614 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:30 crc kubenswrapper[4733]: I1204 19:56:30.330280 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d4qbm" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="registry-server" probeResult="failure" output=< Dec 04 19:56:30 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 19:56:30 crc kubenswrapper[4733]: > Dec 04 19:56:39 crc kubenswrapper[4733]: I1204 19:56:39.362705 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:39 crc kubenswrapper[4733]: I1204 19:56:39.441631 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:39 crc kubenswrapper[4733]: I1204 19:56:39.613293 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4qbm"] Dec 04 19:56:40 crc kubenswrapper[4733]: I1204 19:56:40.531214 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d4qbm" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="registry-server" containerID="cri-o://b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7" gracePeriod=2 Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.136687 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.240954 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-utilities\") pod \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.241123 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-catalog-content\") pod \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.241174 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-645gw\" (UniqueName: \"kubernetes.io/projected/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-kube-api-access-645gw\") pod \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\" (UID: \"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde\") " Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.242485 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-utilities" (OuterVolumeSpecName: "utilities") pod "f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" (UID: "f8097bf8-3766-4c63-b7b3-97cbb7ad0bde"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.248049 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-kube-api-access-645gw" (OuterVolumeSpecName: "kube-api-access-645gw") pod "f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" (UID: "f8097bf8-3766-4c63-b7b3-97cbb7ad0bde"). InnerVolumeSpecName "kube-api-access-645gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.343865 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.343903 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-645gw\" (UniqueName: \"kubernetes.io/projected/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-kube-api-access-645gw\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.360671 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" (UID: "f8097bf8-3766-4c63-b7b3-97cbb7ad0bde"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.446986 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.548829 4733 generic.go:334] "Generic (PLEG): container finished" podID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerID="b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7" exitCode=0 Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.548868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4qbm" event={"ID":"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde","Type":"ContainerDied","Data":"b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7"} Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.548894 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4qbm" event={"ID":"f8097bf8-3766-4c63-b7b3-97cbb7ad0bde","Type":"ContainerDied","Data":"ca382d07398b9b6df253005dcb35286c126caa1d32dfb7e0d39b449405514ffa"} Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.548913 4733 scope.go:117] "RemoveContainer" containerID="b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.548935 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4qbm" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.584624 4733 scope.go:117] "RemoveContainer" containerID="e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.613963 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4qbm"] Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.625332 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d4qbm"] Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.626220 4733 scope.go:117] "RemoveContainer" containerID="0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.675267 4733 scope.go:117] "RemoveContainer" containerID="b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7" Dec 04 19:56:41 crc kubenswrapper[4733]: E1204 19:56:41.675655 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7\": container with ID starting with b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7 not found: ID does not exist" containerID="b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.675700 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7"} err="failed to get container status \"b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7\": rpc error: code = NotFound desc = could not find container \"b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7\": container with ID starting with b7a44cdbfd1a8a9234599257670d39216423c7da143d4d235e42645985404bf7 not found: ID does not exist" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.675734 4733 scope.go:117] "RemoveContainer" containerID="e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9" Dec 04 19:56:41 crc kubenswrapper[4733]: E1204 19:56:41.676215 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9\": container with ID starting with e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9 not found: ID does not exist" containerID="e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.676251 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9"} err="failed to get container status \"e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9\": rpc error: code = NotFound desc = could not find container \"e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9\": container with ID starting with e6ee1afd21cb85d578257882d0bdf88505a3bc5bb99fd05139f86ddb475865f9 not found: ID does not exist" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.676266 4733 scope.go:117] "RemoveContainer" containerID="0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da" Dec 04 19:56:41 crc kubenswrapper[4733]: E1204 19:56:41.676755 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da\": container with ID starting with 0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da not found: ID does not exist" containerID="0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da" Dec 04 19:56:41 crc kubenswrapper[4733]: I1204 19:56:41.676809 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da"} err="failed to get container status \"0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da\": rpc error: code = NotFound desc = could not find container \"0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da\": container with ID starting with 0ddd8cab5e071e4e146241daede1cd5f7ca9ad944c26525495cb7f6747e916da not found: ID does not exist" Dec 04 19:56:42 crc kubenswrapper[4733]: I1204 19:56:42.356430 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" path="/var/lib/kubelet/pods/f8097bf8-3766-4c63-b7b3-97cbb7ad0bde/volumes" Dec 04 19:56:56 crc kubenswrapper[4733]: I1204 19:56:56.721225 4733 generic.go:334] "Generic (PLEG): container finished" podID="df93d53f-da19-405c-a337-bbce7af8fd33" containerID="3730731e9093a3c0f028e1653e5a916fd37ed0f4cba3f31192dce7b4f57c6e7a" exitCode=0 Dec 04 19:56:56 crc kubenswrapper[4733]: I1204 19:56:56.721340 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" event={"ID":"df93d53f-da19-405c-a337-bbce7af8fd33","Type":"ContainerDied","Data":"3730731e9093a3c0f028e1653e5a916fd37ed0f4cba3f31192dce7b4f57c6e7a"} Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.289269 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.349833 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-inventory\") pod \"df93d53f-da19-405c-a337-bbce7af8fd33\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.350048 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-secret-0\") pod \"df93d53f-da19-405c-a337-bbce7af8fd33\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.350095 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ssh-key\") pod \"df93d53f-da19-405c-a337-bbce7af8fd33\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.350280 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ceph\") pod \"df93d53f-da19-405c-a337-bbce7af8fd33\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.350333 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-combined-ca-bundle\") pod \"df93d53f-da19-405c-a337-bbce7af8fd33\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.350438 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmh8z\" (UniqueName: \"kubernetes.io/projected/df93d53f-da19-405c-a337-bbce7af8fd33-kube-api-access-xmh8z\") pod \"df93d53f-da19-405c-a337-bbce7af8fd33\" (UID: \"df93d53f-da19-405c-a337-bbce7af8fd33\") " Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.355539 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df93d53f-da19-405c-a337-bbce7af8fd33-kube-api-access-xmh8z" (OuterVolumeSpecName: "kube-api-access-xmh8z") pod "df93d53f-da19-405c-a337-bbce7af8fd33" (UID: "df93d53f-da19-405c-a337-bbce7af8fd33"). InnerVolumeSpecName "kube-api-access-xmh8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.356526 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ceph" (OuterVolumeSpecName: "ceph") pod "df93d53f-da19-405c-a337-bbce7af8fd33" (UID: "df93d53f-da19-405c-a337-bbce7af8fd33"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.356738 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "df93d53f-da19-405c-a337-bbce7af8fd33" (UID: "df93d53f-da19-405c-a337-bbce7af8fd33"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.385849 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-inventory" (OuterVolumeSpecName: "inventory") pod "df93d53f-da19-405c-a337-bbce7af8fd33" (UID: "df93d53f-da19-405c-a337-bbce7af8fd33"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.410901 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "df93d53f-da19-405c-a337-bbce7af8fd33" (UID: "df93d53f-da19-405c-a337-bbce7af8fd33"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.413962 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "df93d53f-da19-405c-a337-bbce7af8fd33" (UID: "df93d53f-da19-405c-a337-bbce7af8fd33"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.452593 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.452844 4733 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.452939 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmh8z\" (UniqueName: \"kubernetes.io/projected/df93d53f-da19-405c-a337-bbce7af8fd33-kube-api-access-xmh8z\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.453008 4733 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.453095 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.453159 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df93d53f-da19-405c-a337-bbce7af8fd33-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.748903 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" event={"ID":"df93d53f-da19-405c-a337-bbce7af8fd33","Type":"ContainerDied","Data":"cfe295ea748a0690aa6647bb64cb1f67958dc002cbe7e8a92105cb5f6962623c"} Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.749337 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfe295ea748a0690aa6647bb64cb1f67958dc002cbe7e8a92105cb5f6962623c" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.749050 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zk2ps" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.905337 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-lzzhp"] Dec 04 19:56:58 crc kubenswrapper[4733]: E1204 19:56:58.906008 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="extract-content" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.906118 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="extract-content" Dec 04 19:56:58 crc kubenswrapper[4733]: E1204 19:56:58.906201 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="registry-server" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.906278 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="registry-server" Dec 04 19:56:58 crc kubenswrapper[4733]: E1204 19:56:58.906362 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="extract-utilities" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.906427 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="extract-utilities" Dec 04 19:56:58 crc kubenswrapper[4733]: E1204 19:56:58.906522 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df93d53f-da19-405c-a337-bbce7af8fd33" containerName="libvirt-openstack-openstack-cell1" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.906591 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="df93d53f-da19-405c-a337-bbce7af8fd33" containerName="libvirt-openstack-openstack-cell1" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.906917 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="df93d53f-da19-405c-a337-bbce7af8fd33" containerName="libvirt-openstack-openstack-cell1" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.907042 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8097bf8-3766-4c63-b7b3-97cbb7ad0bde" containerName="registry-server" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.908042 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.911513 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.911628 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.911666 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.911882 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.912562 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.912850 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.913496 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.936731 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-lzzhp"] Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.963320 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.963384 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.963470 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.963731 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxljs\" (UniqueName: \"kubernetes.io/projected/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-kube-api-access-fxljs\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.963877 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ceph\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.963986 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.964064 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.964093 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-inventory\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.964131 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.964160 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:58 crc kubenswrapper[4733]: I1204 19:56:58.964238 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.067951 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ceph\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068033 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068076 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068100 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-inventory\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068129 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068148 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068194 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068233 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068259 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068306 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.068378 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxljs\" (UniqueName: \"kubernetes.io/projected/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-kube-api-access-fxljs\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.070135 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.074223 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.075398 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-inventory\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.076166 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.076325 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.077583 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.077900 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ceph\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.078296 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.078409 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.093282 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.107387 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxljs\" (UniqueName: \"kubernetes.io/projected/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-kube-api-access-fxljs\") pod \"nova-cell1-openstack-openstack-cell1-lzzhp\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.238493 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 19:56:59 crc kubenswrapper[4733]: I1204 19:56:59.867282 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-lzzhp"] Dec 04 19:56:59 crc kubenswrapper[4733]: W1204 19:56:59.875938 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97de9284_f0f7_4adf_8f5e_35e3223bd2f8.slice/crio-45ed88817faf4edd5a4b18c19da831acf64a507501edff7843379ea4647b49a9 WatchSource:0}: Error finding container 45ed88817faf4edd5a4b18c19da831acf64a507501edff7843379ea4647b49a9: Status 404 returned error can't find the container with id 45ed88817faf4edd5a4b18c19da831acf64a507501edff7843379ea4647b49a9 Dec 04 19:57:00 crc kubenswrapper[4733]: I1204 19:57:00.775330 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" event={"ID":"97de9284-f0f7-4adf-8f5e-35e3223bd2f8","Type":"ContainerStarted","Data":"45ed88817faf4edd5a4b18c19da831acf64a507501edff7843379ea4647b49a9"} Dec 04 19:57:02 crc kubenswrapper[4733]: I1204 19:57:02.799310 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" event={"ID":"97de9284-f0f7-4adf-8f5e-35e3223bd2f8","Type":"ContainerStarted","Data":"2bb496a6a32e912c09bd905ccfe7a04dd23fa2543d3df6eef0efb218e8498cf8"} Dec 04 19:57:02 crc kubenswrapper[4733]: I1204 19:57:02.829726 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" podStartSLOduration=3.032590125 podStartE2EDuration="4.829709191s" podCreationTimestamp="2025-12-04 19:56:58 +0000 UTC" firstStartedPulling="2025-12-04 19:56:59.878037981 +0000 UTC m=+8281.833399037" lastFinishedPulling="2025-12-04 19:57:01.675157057 +0000 UTC m=+8283.630518103" observedRunningTime="2025-12-04 19:57:02.82265332 +0000 UTC m=+8284.778014376" watchObservedRunningTime="2025-12-04 19:57:02.829709191 +0000 UTC m=+8284.785070237" Dec 04 19:57:15 crc kubenswrapper[4733]: I1204 19:57:15.361923 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:57:15 crc kubenswrapper[4733]: I1204 19:57:15.362573 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:57:45 crc kubenswrapper[4733]: I1204 19:57:45.362336 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:57:45 crc kubenswrapper[4733]: I1204 19:57:45.363254 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:58:15 crc kubenswrapper[4733]: I1204 19:58:15.362567 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 19:58:15 crc kubenswrapper[4733]: I1204 19:58:15.363276 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 19:58:15 crc kubenswrapper[4733]: I1204 19:58:15.363342 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 19:58:15 crc kubenswrapper[4733]: I1204 19:58:15.364585 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06b77c5359e2bbbe0c6023f6861cdc478932b407b029be4bc868ca52a499a232"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 19:58:15 crc kubenswrapper[4733]: I1204 19:58:15.364689 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://06b77c5359e2bbbe0c6023f6861cdc478932b407b029be4bc868ca52a499a232" gracePeriod=600 Dec 04 19:58:15 crc kubenswrapper[4733]: I1204 19:58:15.703836 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="06b77c5359e2bbbe0c6023f6861cdc478932b407b029be4bc868ca52a499a232" exitCode=0 Dec 04 19:58:15 crc kubenswrapper[4733]: I1204 19:58:15.703908 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"06b77c5359e2bbbe0c6023f6861cdc478932b407b029be4bc868ca52a499a232"} Dec 04 19:58:15 crc kubenswrapper[4733]: I1204 19:58:15.704141 4733 scope.go:117] "RemoveContainer" containerID="0247ef920a7d3ae1739541d6a74900a474797ae1a9e55a80f5a835c141e26e5d" Dec 04 19:58:16 crc kubenswrapper[4733]: I1204 19:58:16.722223 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c"} Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.172278 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7"] Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.174430 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.176984 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.177508 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.199272 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7"] Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.261220 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/713695af-62a1-4029-a78a-56014b5d2883-secret-volume\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.261299 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/713695af-62a1-4029-a78a-56014b5d2883-config-volume\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.261343 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxkm4\" (UniqueName: \"kubernetes.io/projected/713695af-62a1-4029-a78a-56014b5d2883-kube-api-access-dxkm4\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.364380 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/713695af-62a1-4029-a78a-56014b5d2883-secret-volume\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.364490 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/713695af-62a1-4029-a78a-56014b5d2883-config-volume\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.364550 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxkm4\" (UniqueName: \"kubernetes.io/projected/713695af-62a1-4029-a78a-56014b5d2883-kube-api-access-dxkm4\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.365918 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/713695af-62a1-4029-a78a-56014b5d2883-config-volume\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.371955 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/713695af-62a1-4029-a78a-56014b5d2883-secret-volume\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.385704 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxkm4\" (UniqueName: \"kubernetes.io/projected/713695af-62a1-4029-a78a-56014b5d2883-kube-api-access-dxkm4\") pod \"collect-profiles-29414640-z4fd7\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:00 crc kubenswrapper[4733]: I1204 20:00:00.502323 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:01 crc kubenswrapper[4733]: I1204 20:00:01.082448 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7"] Dec 04 20:00:01 crc kubenswrapper[4733]: W1204 20:00:01.098703 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod713695af_62a1_4029_a78a_56014b5d2883.slice/crio-186c9e92473ef3f5d24c8df239030ed803edbfcea7d8239b1ee4c9c5a8bfdd51 WatchSource:0}: Error finding container 186c9e92473ef3f5d24c8df239030ed803edbfcea7d8239b1ee4c9c5a8bfdd51: Status 404 returned error can't find the container with id 186c9e92473ef3f5d24c8df239030ed803edbfcea7d8239b1ee4c9c5a8bfdd51 Dec 04 20:00:01 crc kubenswrapper[4733]: I1204 20:00:01.967043 4733 generic.go:334] "Generic (PLEG): container finished" podID="713695af-62a1-4029-a78a-56014b5d2883" containerID="c8a0e22348a9aa7c1f2f461b23d73727e6b0832d7bf3152652ff2e04cdb87c8d" exitCode=0 Dec 04 20:00:01 crc kubenswrapper[4733]: I1204 20:00:01.967107 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" event={"ID":"713695af-62a1-4029-a78a-56014b5d2883","Type":"ContainerDied","Data":"c8a0e22348a9aa7c1f2f461b23d73727e6b0832d7bf3152652ff2e04cdb87c8d"} Dec 04 20:00:01 crc kubenswrapper[4733]: I1204 20:00:01.967324 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" event={"ID":"713695af-62a1-4029-a78a-56014b5d2883","Type":"ContainerStarted","Data":"186c9e92473ef3f5d24c8df239030ed803edbfcea7d8239b1ee4c9c5a8bfdd51"} Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.441383 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.537047 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/713695af-62a1-4029-a78a-56014b5d2883-secret-volume\") pod \"713695af-62a1-4029-a78a-56014b5d2883\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.537304 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxkm4\" (UniqueName: \"kubernetes.io/projected/713695af-62a1-4029-a78a-56014b5d2883-kube-api-access-dxkm4\") pod \"713695af-62a1-4029-a78a-56014b5d2883\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.537558 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/713695af-62a1-4029-a78a-56014b5d2883-config-volume\") pod \"713695af-62a1-4029-a78a-56014b5d2883\" (UID: \"713695af-62a1-4029-a78a-56014b5d2883\") " Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.538258 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/713695af-62a1-4029-a78a-56014b5d2883-config-volume" (OuterVolumeSpecName: "config-volume") pod "713695af-62a1-4029-a78a-56014b5d2883" (UID: "713695af-62a1-4029-a78a-56014b5d2883"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.544991 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/713695af-62a1-4029-a78a-56014b5d2883-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "713695af-62a1-4029-a78a-56014b5d2883" (UID: "713695af-62a1-4029-a78a-56014b5d2883"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.545093 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/713695af-62a1-4029-a78a-56014b5d2883-kube-api-access-dxkm4" (OuterVolumeSpecName: "kube-api-access-dxkm4") pod "713695af-62a1-4029-a78a-56014b5d2883" (UID: "713695af-62a1-4029-a78a-56014b5d2883"). InnerVolumeSpecName "kube-api-access-dxkm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.640207 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/713695af-62a1-4029-a78a-56014b5d2883-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.640257 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxkm4\" (UniqueName: \"kubernetes.io/projected/713695af-62a1-4029-a78a-56014b5d2883-kube-api-access-dxkm4\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.640277 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/713695af-62a1-4029-a78a-56014b5d2883-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.999129 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" event={"ID":"713695af-62a1-4029-a78a-56014b5d2883","Type":"ContainerDied","Data":"186c9e92473ef3f5d24c8df239030ed803edbfcea7d8239b1ee4c9c5a8bfdd51"} Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.999189 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="186c9e92473ef3f5d24c8df239030ed803edbfcea7d8239b1ee4c9c5a8bfdd51" Dec 04 20:00:03 crc kubenswrapper[4733]: I1204 20:00:03.999208 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414640-z4fd7" Dec 04 20:00:04 crc kubenswrapper[4733]: I1204 20:00:04.540529 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s"] Dec 04 20:00:04 crc kubenswrapper[4733]: I1204 20:00:04.551531 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414595-v2f6s"] Dec 04 20:00:06 crc kubenswrapper[4733]: I1204 20:00:06.351780 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3ac4c48-5b84-4803-a16d-c3c96effa7f5" path="/var/lib/kubelet/pods/f3ac4c48-5b84-4803-a16d-c3c96effa7f5/volumes" Dec 04 20:00:15 crc kubenswrapper[4733]: I1204 20:00:15.361579 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:00:15 crc kubenswrapper[4733]: I1204 20:00:15.362299 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:00:18 crc kubenswrapper[4733]: I1204 20:00:18.164575 4733 generic.go:334] "Generic (PLEG): container finished" podID="97de9284-f0f7-4adf-8f5e-35e3223bd2f8" containerID="2bb496a6a32e912c09bd905ccfe7a04dd23fa2543d3df6eef0efb218e8498cf8" exitCode=0 Dec 04 20:00:18 crc kubenswrapper[4733]: I1204 20:00:18.164686 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" event={"ID":"97de9284-f0f7-4adf-8f5e-35e3223bd2f8","Type":"ContainerDied","Data":"2bb496a6a32e912c09bd905ccfe7a04dd23fa2543d3df6eef0efb218e8498cf8"} Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.659048 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.708601 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-combined-ca-bundle\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.708848 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-1\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.708922 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxljs\" (UniqueName: \"kubernetes.io/projected/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-kube-api-access-fxljs\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.709088 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-0\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.709174 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-inventory\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.709225 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ssh-key\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.709273 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ceph\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.709332 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-0\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.709392 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-1\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.709445 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-1\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.709550 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-0\") pod \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\" (UID: \"97de9284-f0f7-4adf-8f5e-35e3223bd2f8\") " Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.730371 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ceph" (OuterVolumeSpecName: "ceph") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.732306 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-kube-api-access-fxljs" (OuterVolumeSpecName: "kube-api-access-fxljs") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "kube-api-access-fxljs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.751756 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.766083 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.772397 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.776226 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-inventory" (OuterVolumeSpecName: "inventory") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.777961 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.788632 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.791774 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.794279 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.803849 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "97de9284-f0f7-4adf-8f5e-35e3223bd2f8" (UID: "97de9284-f0f7-4adf-8f5e-35e3223bd2f8"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812745 4733 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812773 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812783 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812806 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxljs\" (UniqueName: \"kubernetes.io/projected/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-kube-api-access-fxljs\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812819 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812830 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812838 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812847 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812855 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812863 4733 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:19 crc kubenswrapper[4733]: I1204 20:00:19.812871 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/97de9284-f0f7-4adf-8f5e-35e3223bd2f8-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.190169 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" event={"ID":"97de9284-f0f7-4adf-8f5e-35e3223bd2f8","Type":"ContainerDied","Data":"45ed88817faf4edd5a4b18c19da831acf64a507501edff7843379ea4647b49a9"} Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.190500 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45ed88817faf4edd5a4b18c19da831acf64a507501edff7843379ea4647b49a9" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.190237 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-lzzhp" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.311460 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-hhgzn"] Dec 04 20:00:20 crc kubenswrapper[4733]: E1204 20:00:20.311883 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="713695af-62a1-4029-a78a-56014b5d2883" containerName="collect-profiles" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.311899 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="713695af-62a1-4029-a78a-56014b5d2883" containerName="collect-profiles" Dec 04 20:00:20 crc kubenswrapper[4733]: E1204 20:00:20.311919 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97de9284-f0f7-4adf-8f5e-35e3223bd2f8" containerName="nova-cell1-openstack-openstack-cell1" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.311927 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="97de9284-f0f7-4adf-8f5e-35e3223bd2f8" containerName="nova-cell1-openstack-openstack-cell1" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.312112 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="713695af-62a1-4029-a78a-56014b5d2883" containerName="collect-profiles" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.312139 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="97de9284-f0f7-4adf-8f5e-35e3223bd2f8" containerName="nova-cell1-openstack-openstack-cell1" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.312811 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.318194 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.318418 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.318357 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.318853 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.318991 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.334528 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-hhgzn"] Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.427830 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.427994 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceph\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.428064 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.428241 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hp5s\" (UniqueName: \"kubernetes.io/projected/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-kube-api-access-2hp5s\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.428409 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ssh-key\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.429125 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-inventory\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.429252 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.429296 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.531452 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.531533 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.531574 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.531601 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceph\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.531622 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.531654 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hp5s\" (UniqueName: \"kubernetes.io/projected/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-kube-api-access-2hp5s\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.531696 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ssh-key\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.531862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-inventory\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.536533 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.537437 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceph\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.537558 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ssh-key\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.537888 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.538187 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.538263 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.548867 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-inventory\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.551608 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hp5s\" (UniqueName: \"kubernetes.io/projected/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-kube-api-access-2hp5s\") pod \"telemetry-openstack-openstack-cell1-hhgzn\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:20 crc kubenswrapper[4733]: I1204 20:00:20.632575 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:00:21 crc kubenswrapper[4733]: I1204 20:00:21.346433 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-hhgzn"] Dec 04 20:00:21 crc kubenswrapper[4733]: I1204 20:00:21.347579 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 20:00:22 crc kubenswrapper[4733]: I1204 20:00:22.214124 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" event={"ID":"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed","Type":"ContainerStarted","Data":"f22abcd789bcc50104284e6979ab1da517b173d4b4b78316eed1e3feeb8a445b"} Dec 04 20:00:22 crc kubenswrapper[4733]: I1204 20:00:22.214482 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" event={"ID":"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed","Type":"ContainerStarted","Data":"55ba7fc6018808c748f8c3d852b7c89932ab8e70c464473d2405b40addc02a1e"} Dec 04 20:00:22 crc kubenswrapper[4733]: I1204 20:00:22.240946 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" podStartSLOduration=1.776613888 podStartE2EDuration="2.240924598s" podCreationTimestamp="2025-12-04 20:00:20 +0000 UTC" firstStartedPulling="2025-12-04 20:00:21.347165265 +0000 UTC m=+8483.302526311" lastFinishedPulling="2025-12-04 20:00:21.811475935 +0000 UTC m=+8483.766837021" observedRunningTime="2025-12-04 20:00:22.233771715 +0000 UTC m=+8484.189132781" watchObservedRunningTime="2025-12-04 20:00:22.240924598 +0000 UTC m=+8484.196285654" Dec 04 20:00:28 crc kubenswrapper[4733]: I1204 20:00:28.933012 4733 scope.go:117] "RemoveContainer" containerID="54a484384f97297c17ae390955e4d4e3a9e8d9c5381d7df03e915d2ba526887f" Dec 04 20:00:45 crc kubenswrapper[4733]: I1204 20:00:45.362140 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:00:45 crc kubenswrapper[4733]: I1204 20:00:45.362876 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.163432 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29414641-j7k8n"] Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.165731 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.197560 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414641-j7k8n"] Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.331006 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-combined-ca-bundle\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.331165 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-config-data\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.331280 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-fernet-keys\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.331349 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtnj2\" (UniqueName: \"kubernetes.io/projected/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-kube-api-access-qtnj2\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.433141 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-fernet-keys\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.433228 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtnj2\" (UniqueName: \"kubernetes.io/projected/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-kube-api-access-qtnj2\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.433284 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-combined-ca-bundle\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.433381 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-config-data\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.439117 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-config-data\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.439382 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-fernet-keys\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.439623 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-combined-ca-bundle\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.452516 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtnj2\" (UniqueName: \"kubernetes.io/projected/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-kube-api-access-qtnj2\") pod \"keystone-cron-29414641-j7k8n\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.496051 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:00 crc kubenswrapper[4733]: I1204 20:01:00.990431 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414641-j7k8n"] Dec 04 20:01:01 crc kubenswrapper[4733]: I1204 20:01:01.689671 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414641-j7k8n" event={"ID":"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8","Type":"ContainerStarted","Data":"6c772ad82dd93f60633715a5445b16b2de8af6c4f731f116f9b4d1fb3e0cf617"} Dec 04 20:01:01 crc kubenswrapper[4733]: I1204 20:01:01.689998 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414641-j7k8n" event={"ID":"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8","Type":"ContainerStarted","Data":"5278a63df5f9c68c478c54f6065af5af32c855816b3e6501212ab762ebd29516"} Dec 04 20:01:01 crc kubenswrapper[4733]: I1204 20:01:01.709823 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29414641-j7k8n" podStartSLOduration=1.709788762 podStartE2EDuration="1.709788762s" podCreationTimestamp="2025-12-04 20:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 20:01:01.704310253 +0000 UTC m=+8523.659671309" watchObservedRunningTime="2025-12-04 20:01:01.709788762 +0000 UTC m=+8523.665149808" Dec 04 20:01:03 crc kubenswrapper[4733]: I1204 20:01:03.713934 4733 generic.go:334] "Generic (PLEG): container finished" podID="e83b39f0-0ffc-46d2-83b8-fbccf10aeef8" containerID="6c772ad82dd93f60633715a5445b16b2de8af6c4f731f116f9b4d1fb3e0cf617" exitCode=0 Dec 04 20:01:03 crc kubenswrapper[4733]: I1204 20:01:03.714069 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414641-j7k8n" event={"ID":"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8","Type":"ContainerDied","Data":"6c772ad82dd93f60633715a5445b16b2de8af6c4f731f116f9b4d1fb3e0cf617"} Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.135950 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.223256 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-combined-ca-bundle\") pod \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.223334 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-config-data\") pod \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.223411 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtnj2\" (UniqueName: \"kubernetes.io/projected/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-kube-api-access-qtnj2\") pod \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.223469 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-fernet-keys\") pod \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\" (UID: \"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8\") " Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.228741 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e83b39f0-0ffc-46d2-83b8-fbccf10aeef8" (UID: "e83b39f0-0ffc-46d2-83b8-fbccf10aeef8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.232888 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-kube-api-access-qtnj2" (OuterVolumeSpecName: "kube-api-access-qtnj2") pod "e83b39f0-0ffc-46d2-83b8-fbccf10aeef8" (UID: "e83b39f0-0ffc-46d2-83b8-fbccf10aeef8"). InnerVolumeSpecName "kube-api-access-qtnj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.258634 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e83b39f0-0ffc-46d2-83b8-fbccf10aeef8" (UID: "e83b39f0-0ffc-46d2-83b8-fbccf10aeef8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.278203 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-config-data" (OuterVolumeSpecName: "config-data") pod "e83b39f0-0ffc-46d2-83b8-fbccf10aeef8" (UID: "e83b39f0-0ffc-46d2-83b8-fbccf10aeef8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.326262 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.326297 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtnj2\" (UniqueName: \"kubernetes.io/projected/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-kube-api-access-qtnj2\") on node \"crc\" DevicePath \"\"" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.326307 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.326316 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83b39f0-0ffc-46d2-83b8-fbccf10aeef8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.741658 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414641-j7k8n" event={"ID":"e83b39f0-0ffc-46d2-83b8-fbccf10aeef8","Type":"ContainerDied","Data":"5278a63df5f9c68c478c54f6065af5af32c855816b3e6501212ab762ebd29516"} Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.741700 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5278a63df5f9c68c478c54f6065af5af32c855816b3e6501212ab762ebd29516" Dec 04 20:01:05 crc kubenswrapper[4733]: I1204 20:01:05.741708 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414641-j7k8n" Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.362050 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.362712 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.362768 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.364297 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.364436 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" gracePeriod=600 Dec 04 20:01:15 crc kubenswrapper[4733]: E1204 20:01:15.502836 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.866076 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" exitCode=0 Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.866145 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c"} Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.866551 4733 scope.go:117] "RemoveContainer" containerID="06b77c5359e2bbbe0c6023f6861cdc478932b407b029be4bc868ca52a499a232" Dec 04 20:01:15 crc kubenswrapper[4733]: I1204 20:01:15.867351 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:01:15 crc kubenswrapper[4733]: E1204 20:01:15.867654 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:01:29 crc kubenswrapper[4733]: I1204 20:01:29.337899 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:01:29 crc kubenswrapper[4733]: E1204 20:01:29.338881 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:01:41 crc kubenswrapper[4733]: I1204 20:01:41.335659 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:01:41 crc kubenswrapper[4733]: E1204 20:01:41.336490 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:01:55 crc kubenswrapper[4733]: I1204 20:01:55.335648 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:01:55 crc kubenswrapper[4733]: E1204 20:01:55.336834 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:02:09 crc kubenswrapper[4733]: I1204 20:02:09.336537 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:02:09 crc kubenswrapper[4733]: E1204 20:02:09.337586 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:02:20 crc kubenswrapper[4733]: I1204 20:02:20.336735 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:02:20 crc kubenswrapper[4733]: E1204 20:02:20.337815 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.415374 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-66628"] Dec 04 20:02:28 crc kubenswrapper[4733]: E1204 20:02:28.416405 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83b39f0-0ffc-46d2-83b8-fbccf10aeef8" containerName="keystone-cron" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.416417 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83b39f0-0ffc-46d2-83b8-fbccf10aeef8" containerName="keystone-cron" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.416614 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83b39f0-0ffc-46d2-83b8-fbccf10aeef8" containerName="keystone-cron" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.418428 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.426574 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-66628"] Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.608742 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-utilities\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.609164 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-catalog-content\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.609209 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh26r\" (UniqueName: \"kubernetes.io/projected/2155a180-bd8f-4fd0-96c4-b173e0e213cf-kube-api-access-qh26r\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.710923 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-utilities\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.711014 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-catalog-content\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.711058 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh26r\" (UniqueName: \"kubernetes.io/projected/2155a180-bd8f-4fd0-96c4-b173e0e213cf-kube-api-access-qh26r\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.711513 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-utilities\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.711545 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-catalog-content\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.739357 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh26r\" (UniqueName: \"kubernetes.io/projected/2155a180-bd8f-4fd0-96c4-b173e0e213cf-kube-api-access-qh26r\") pod \"certified-operators-66628\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:28 crc kubenswrapper[4733]: I1204 20:02:28.804815 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:29 crc kubenswrapper[4733]: I1204 20:02:29.401262 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-66628"] Dec 04 20:02:29 crc kubenswrapper[4733]: I1204 20:02:29.804515 4733 generic.go:334] "Generic (PLEG): container finished" podID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerID="efa4771cc4824329132444d78e03abbb67d6577ef7b3a69256616bde3dedae06" exitCode=0 Dec 04 20:02:29 crc kubenswrapper[4733]: I1204 20:02:29.804665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66628" event={"ID":"2155a180-bd8f-4fd0-96c4-b173e0e213cf","Type":"ContainerDied","Data":"efa4771cc4824329132444d78e03abbb67d6577ef7b3a69256616bde3dedae06"} Dec 04 20:02:29 crc kubenswrapper[4733]: I1204 20:02:29.804870 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66628" event={"ID":"2155a180-bd8f-4fd0-96c4-b173e0e213cf","Type":"ContainerStarted","Data":"4d454ecd79c595d84ca3618bd934d8284eac1a5f137b7ba5bf2bc58c0048e211"} Dec 04 20:02:30 crc kubenswrapper[4733]: I1204 20:02:30.821112 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66628" event={"ID":"2155a180-bd8f-4fd0-96c4-b173e0e213cf","Type":"ContainerStarted","Data":"a8bbf759b1b623af21418f0018e669ee2209b9a805107c6e81c128848da38ccd"} Dec 04 20:02:31 crc kubenswrapper[4733]: I1204 20:02:31.336079 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:02:31 crc kubenswrapper[4733]: E1204 20:02:31.336858 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:02:31 crc kubenswrapper[4733]: I1204 20:02:31.841640 4733 generic.go:334] "Generic (PLEG): container finished" podID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerID="a8bbf759b1b623af21418f0018e669ee2209b9a805107c6e81c128848da38ccd" exitCode=0 Dec 04 20:02:31 crc kubenswrapper[4733]: I1204 20:02:31.841750 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66628" event={"ID":"2155a180-bd8f-4fd0-96c4-b173e0e213cf","Type":"ContainerDied","Data":"a8bbf759b1b623af21418f0018e669ee2209b9a805107c6e81c128848da38ccd"} Dec 04 20:02:32 crc kubenswrapper[4733]: I1204 20:02:32.855518 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66628" event={"ID":"2155a180-bd8f-4fd0-96c4-b173e0e213cf","Type":"ContainerStarted","Data":"5b45936599b2a51ceaad1637192107cdb6d8abc7970068d5a905f493dbd1b783"} Dec 04 20:02:32 crc kubenswrapper[4733]: I1204 20:02:32.880236 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-66628" podStartSLOduration=2.439075104 podStartE2EDuration="4.880211224s" podCreationTimestamp="2025-12-04 20:02:28 +0000 UTC" firstStartedPulling="2025-12-04 20:02:29.806565766 +0000 UTC m=+8611.761926812" lastFinishedPulling="2025-12-04 20:02:32.247701876 +0000 UTC m=+8614.203062932" observedRunningTime="2025-12-04 20:02:32.873271617 +0000 UTC m=+8614.828632683" watchObservedRunningTime="2025-12-04 20:02:32.880211224 +0000 UTC m=+8614.835572270" Dec 04 20:02:38 crc kubenswrapper[4733]: I1204 20:02:38.806410 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:38 crc kubenswrapper[4733]: I1204 20:02:38.807021 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:38 crc kubenswrapper[4733]: I1204 20:02:38.906157 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:39 crc kubenswrapper[4733]: I1204 20:02:39.008472 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:39 crc kubenswrapper[4733]: I1204 20:02:39.148684 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-66628"] Dec 04 20:02:40 crc kubenswrapper[4733]: I1204 20:02:40.964177 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-66628" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerName="registry-server" containerID="cri-o://5b45936599b2a51ceaad1637192107cdb6d8abc7970068d5a905f493dbd1b783" gracePeriod=2 Dec 04 20:02:41 crc kubenswrapper[4733]: E1204 20:02:41.232777 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2155a180_bd8f_4fd0_96c4_b173e0e213cf.slice/crio-5b45936599b2a51ceaad1637192107cdb6d8abc7970068d5a905f493dbd1b783.scope\": RecentStats: unable to find data in memory cache]" Dec 04 20:02:41 crc kubenswrapper[4733]: I1204 20:02:41.981576 4733 generic.go:334] "Generic (PLEG): container finished" podID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerID="5b45936599b2a51ceaad1637192107cdb6d8abc7970068d5a905f493dbd1b783" exitCode=0 Dec 04 20:02:41 crc kubenswrapper[4733]: I1204 20:02:41.981620 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66628" event={"ID":"2155a180-bd8f-4fd0-96c4-b173e0e213cf","Type":"ContainerDied","Data":"5b45936599b2a51ceaad1637192107cdb6d8abc7970068d5a905f493dbd1b783"} Dec 04 20:02:41 crc kubenswrapper[4733]: I1204 20:02:41.981939 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66628" event={"ID":"2155a180-bd8f-4fd0-96c4-b173e0e213cf","Type":"ContainerDied","Data":"4d454ecd79c595d84ca3618bd934d8284eac1a5f137b7ba5bf2bc58c0048e211"} Dec 04 20:02:41 crc kubenswrapper[4733]: I1204 20:02:41.981970 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d454ecd79c595d84ca3618bd934d8284eac1a5f137b7ba5bf2bc58c0048e211" Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.091501 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.240852 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh26r\" (UniqueName: \"kubernetes.io/projected/2155a180-bd8f-4fd0-96c4-b173e0e213cf-kube-api-access-qh26r\") pod \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.241032 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-utilities\") pod \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.241081 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-catalog-content\") pod \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\" (UID: \"2155a180-bd8f-4fd0-96c4-b173e0e213cf\") " Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.254142 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-utilities" (OuterVolumeSpecName: "utilities") pod "2155a180-bd8f-4fd0-96c4-b173e0e213cf" (UID: "2155a180-bd8f-4fd0-96c4-b173e0e213cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.260261 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2155a180-bd8f-4fd0-96c4-b173e0e213cf-kube-api-access-qh26r" (OuterVolumeSpecName: "kube-api-access-qh26r") pod "2155a180-bd8f-4fd0-96c4-b173e0e213cf" (UID: "2155a180-bd8f-4fd0-96c4-b173e0e213cf"). InnerVolumeSpecName "kube-api-access-qh26r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.291868 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2155a180-bd8f-4fd0-96c4-b173e0e213cf" (UID: "2155a180-bd8f-4fd0-96c4-b173e0e213cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.343112 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh26r\" (UniqueName: \"kubernetes.io/projected/2155a180-bd8f-4fd0-96c4-b173e0e213cf-kube-api-access-qh26r\") on node \"crc\" DevicePath \"\"" Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.343144 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.343153 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2155a180-bd8f-4fd0-96c4-b173e0e213cf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:02:42 crc kubenswrapper[4733]: I1204 20:02:42.993065 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-66628" Dec 04 20:02:43 crc kubenswrapper[4733]: I1204 20:02:43.031415 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-66628"] Dec 04 20:02:43 crc kubenswrapper[4733]: I1204 20:02:43.041633 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-66628"] Dec 04 20:02:44 crc kubenswrapper[4733]: I1204 20:02:44.353990 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" path="/var/lib/kubelet/pods/2155a180-bd8f-4fd0-96c4-b173e0e213cf/volumes" Dec 04 20:02:46 crc kubenswrapper[4733]: I1204 20:02:46.337093 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:02:46 crc kubenswrapper[4733]: E1204 20:02:46.337580 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:02:57 crc kubenswrapper[4733]: I1204 20:02:57.336206 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:02:57 crc kubenswrapper[4733]: E1204 20:02:57.337957 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:03:09 crc kubenswrapper[4733]: I1204 20:03:09.336109 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:03:09 crc kubenswrapper[4733]: E1204 20:03:09.337237 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:03:21 crc kubenswrapper[4733]: I1204 20:03:21.336900 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:03:21 crc kubenswrapper[4733]: E1204 20:03:21.337819 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:03:33 crc kubenswrapper[4733]: I1204 20:03:33.336080 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:03:33 crc kubenswrapper[4733]: E1204 20:03:33.337545 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:03:47 crc kubenswrapper[4733]: I1204 20:03:47.337208 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:03:47 crc kubenswrapper[4733]: E1204 20:03:47.338219 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:03:59 crc kubenswrapper[4733]: I1204 20:03:59.336690 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:03:59 crc kubenswrapper[4733]: E1204 20:03:59.337846 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:04:14 crc kubenswrapper[4733]: I1204 20:04:14.336769 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:04:14 crc kubenswrapper[4733]: E1204 20:04:14.337859 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:04:26 crc kubenswrapper[4733]: I1204 20:04:26.336477 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:04:26 crc kubenswrapper[4733]: E1204 20:04:26.337451 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:04:37 crc kubenswrapper[4733]: I1204 20:04:37.336153 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:04:37 crc kubenswrapper[4733]: E1204 20:04:37.337080 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:04:51 crc kubenswrapper[4733]: I1204 20:04:51.335379 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:04:51 crc kubenswrapper[4733]: E1204 20:04:51.336268 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:05:01 crc kubenswrapper[4733]: I1204 20:05:01.627094 4733 generic.go:334] "Generic (PLEG): container finished" podID="469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" containerID="f22abcd789bcc50104284e6979ab1da517b173d4b4b78316eed1e3feeb8a445b" exitCode=0 Dec 04 20:05:01 crc kubenswrapper[4733]: I1204 20:05:01.627176 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" event={"ID":"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed","Type":"ContainerDied","Data":"f22abcd789bcc50104284e6979ab1da517b173d4b4b78316eed1e3feeb8a445b"} Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.154613 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.255629 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-inventory\") pod \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.255703 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-0\") pod \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.255724 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ssh-key\") pod \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.255753 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hp5s\" (UniqueName: \"kubernetes.io/projected/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-kube-api-access-2hp5s\") pod \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.255784 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-1\") pod \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.255910 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceph\") pod \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.255955 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-telemetry-combined-ca-bundle\") pod \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.256002 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-2\") pod \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\" (UID: \"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed\") " Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.262299 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceph" (OuterVolumeSpecName: "ceph") pod "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" (UID: "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.262766 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" (UID: "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.267074 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-kube-api-access-2hp5s" (OuterVolumeSpecName: "kube-api-access-2hp5s") pod "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" (UID: "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed"). InnerVolumeSpecName "kube-api-access-2hp5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.295656 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" (UID: "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.296202 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" (UID: "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.298151 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" (UID: "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.298636 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" (UID: "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.301376 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-inventory" (OuterVolumeSpecName: "inventory") pod "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" (UID: "469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.359034 4733 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.359115 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.359137 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hp5s\" (UniqueName: \"kubernetes.io/projected/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-kube-api-access-2hp5s\") on node \"crc\" DevicePath \"\"" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.359158 4733 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.359175 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.359191 4733 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.359208 4733 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.359224 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.651790 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" event={"ID":"469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed","Type":"ContainerDied","Data":"55ba7fc6018808c748f8c3d852b7c89932ab8e70c464473d2405b40addc02a1e"} Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.652659 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55ba7fc6018808c748f8c3d852b7c89932ab8e70c464473d2405b40addc02a1e" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.651853 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-hhgzn" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.768161 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-jgtmh"] Dec 04 20:05:03 crc kubenswrapper[4733]: E1204 20:05:03.768641 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerName="registry-server" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.768663 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerName="registry-server" Dec 04 20:05:03 crc kubenswrapper[4733]: E1204 20:05:03.768707 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerName="extract-utilities" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.768714 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerName="extract-utilities" Dec 04 20:05:03 crc kubenswrapper[4733]: E1204 20:05:03.768725 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerName="extract-content" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.768732 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerName="extract-content" Dec 04 20:05:03 crc kubenswrapper[4733]: E1204 20:05:03.768747 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" containerName="telemetry-openstack-openstack-cell1" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.768753 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" containerName="telemetry-openstack-openstack-cell1" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.768973 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed" containerName="telemetry-openstack-openstack-cell1" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.768984 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2155a180-bd8f-4fd0-96c4-b173e0e213cf" containerName="registry-server" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.769733 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.772647 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.772864 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.773105 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.773107 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.776117 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.794740 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-jgtmh"] Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.879138 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.879437 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.879537 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.879619 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhwh5\" (UniqueName: \"kubernetes.io/projected/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-kube-api-access-xhwh5\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.879727 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.879879 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.982538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.982898 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.983069 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.983303 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.983443 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.983553 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhwh5\" (UniqueName: \"kubernetes.io/projected/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-kube-api-access-xhwh5\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.989487 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.990760 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.992049 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.993322 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.995885 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:03 crc kubenswrapper[4733]: I1204 20:05:03.999448 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhwh5\" (UniqueName: \"kubernetes.io/projected/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-kube-api-access-xhwh5\") pod \"neutron-sriov-openstack-openstack-cell1-jgtmh\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:04 crc kubenswrapper[4733]: I1204 20:05:04.088209 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:05:04 crc kubenswrapper[4733]: I1204 20:05:04.335758 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:05:04 crc kubenswrapper[4733]: E1204 20:05:04.336485 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:05:04 crc kubenswrapper[4733]: I1204 20:05:04.695973 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-jgtmh"] Dec 04 20:05:05 crc kubenswrapper[4733]: I1204 20:05:05.676689 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" event={"ID":"af2a17da-c85e-42d9-ab67-5c44b5c5eab8","Type":"ContainerStarted","Data":"fd9f2a61ffecf02719349b6b59db8af58812bb83d1e8adfd48ec26c02b7da01a"} Dec 04 20:05:05 crc kubenswrapper[4733]: I1204 20:05:05.677240 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" event={"ID":"af2a17da-c85e-42d9-ab67-5c44b5c5eab8","Type":"ContainerStarted","Data":"84b2d9a0e800a66fafe94424ef1002e0500a35aaaf8127210e3e6f7ca5fe2479"} Dec 04 20:05:05 crc kubenswrapper[4733]: I1204 20:05:05.697914 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" podStartSLOduration=2.182517003 podStartE2EDuration="2.6978914s" podCreationTimestamp="2025-12-04 20:05:03 +0000 UTC" firstStartedPulling="2025-12-04 20:05:04.69722361 +0000 UTC m=+8766.652584666" lastFinishedPulling="2025-12-04 20:05:05.212597997 +0000 UTC m=+8767.167959063" observedRunningTime="2025-12-04 20:05:05.695477064 +0000 UTC m=+8767.650838110" watchObservedRunningTime="2025-12-04 20:05:05.6978914 +0000 UTC m=+8767.653252446" Dec 04 20:05:17 crc kubenswrapper[4733]: I1204 20:05:17.341101 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:05:17 crc kubenswrapper[4733]: E1204 20:05:17.341943 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:05:32 crc kubenswrapper[4733]: I1204 20:05:32.336984 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:05:32 crc kubenswrapper[4733]: E1204 20:05:32.337677 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:05:45 crc kubenswrapper[4733]: I1204 20:05:45.336865 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:05:45 crc kubenswrapper[4733]: E1204 20:05:45.337656 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:05:56 crc kubenswrapper[4733]: I1204 20:05:56.336018 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:05:56 crc kubenswrapper[4733]: E1204 20:05:56.337070 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.051529 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kx8z2"] Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.059612 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.074630 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx8z2"] Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.094635 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-utilities\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.094836 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-catalog-content\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.094897 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59tts\" (UniqueName: \"kubernetes.io/projected/08706f9e-bd19-4f02-9d9c-27a218ec6764-kube-api-access-59tts\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.196363 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-utilities\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.196635 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-catalog-content\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.196725 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59tts\" (UniqueName: \"kubernetes.io/projected/08706f9e-bd19-4f02-9d9c-27a218ec6764-kube-api-access-59tts\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.197334 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-utilities\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.197547 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-catalog-content\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.217938 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59tts\" (UniqueName: \"kubernetes.io/projected/08706f9e-bd19-4f02-9d9c-27a218ec6764-kube-api-access-59tts\") pod \"redhat-marketplace-kx8z2\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.400030 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:01 crc kubenswrapper[4733]: I1204 20:06:01.924981 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx8z2"] Dec 04 20:06:02 crc kubenswrapper[4733]: I1204 20:06:02.936782 4733 generic.go:334] "Generic (PLEG): container finished" podID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerID="91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3" exitCode=0 Dec 04 20:06:02 crc kubenswrapper[4733]: I1204 20:06:02.937026 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx8z2" event={"ID":"08706f9e-bd19-4f02-9d9c-27a218ec6764","Type":"ContainerDied","Data":"91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3"} Dec 04 20:06:02 crc kubenswrapper[4733]: I1204 20:06:02.937102 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx8z2" event={"ID":"08706f9e-bd19-4f02-9d9c-27a218ec6764","Type":"ContainerStarted","Data":"73c9a7dcde4d1222b3692990e2af772e7265983e1f61c1c0eebb15431eebde3e"} Dec 04 20:06:02 crc kubenswrapper[4733]: I1204 20:06:02.941353 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 20:06:03 crc kubenswrapper[4733]: I1204 20:06:03.950866 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx8z2" event={"ID":"08706f9e-bd19-4f02-9d9c-27a218ec6764","Type":"ContainerStarted","Data":"39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e"} Dec 04 20:06:04 crc kubenswrapper[4733]: I1204 20:06:04.965948 4733 generic.go:334] "Generic (PLEG): container finished" podID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerID="39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e" exitCode=0 Dec 04 20:06:04 crc kubenswrapper[4733]: I1204 20:06:04.965999 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx8z2" event={"ID":"08706f9e-bd19-4f02-9d9c-27a218ec6764","Type":"ContainerDied","Data":"39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e"} Dec 04 20:06:05 crc kubenswrapper[4733]: I1204 20:06:05.983652 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx8z2" event={"ID":"08706f9e-bd19-4f02-9d9c-27a218ec6764","Type":"ContainerStarted","Data":"58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8"} Dec 04 20:06:06 crc kubenswrapper[4733]: I1204 20:06:06.010986 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kx8z2" podStartSLOduration=2.543920786 podStartE2EDuration="5.010968407s" podCreationTimestamp="2025-12-04 20:06:01 +0000 UTC" firstStartedPulling="2025-12-04 20:06:02.94113895 +0000 UTC m=+8824.896499986" lastFinishedPulling="2025-12-04 20:06:05.408186561 +0000 UTC m=+8827.363547607" observedRunningTime="2025-12-04 20:06:06.006555207 +0000 UTC m=+8827.961916263" watchObservedRunningTime="2025-12-04 20:06:06.010968407 +0000 UTC m=+8827.966329453" Dec 04 20:06:09 crc kubenswrapper[4733]: I1204 20:06:09.336179 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:06:09 crc kubenswrapper[4733]: E1204 20:06:09.337151 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:06:11 crc kubenswrapper[4733]: I1204 20:06:11.404066 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:11 crc kubenswrapper[4733]: I1204 20:06:11.405561 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:11 crc kubenswrapper[4733]: I1204 20:06:11.460671 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:12 crc kubenswrapper[4733]: I1204 20:06:12.101449 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:12 crc kubenswrapper[4733]: I1204 20:06:12.162765 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx8z2"] Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.071521 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kx8z2" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerName="registry-server" containerID="cri-o://58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8" gracePeriod=2 Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.578417 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.709459 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-catalog-content\") pod \"08706f9e-bd19-4f02-9d9c-27a218ec6764\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.709929 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-utilities\") pod \"08706f9e-bd19-4f02-9d9c-27a218ec6764\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.709992 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59tts\" (UniqueName: \"kubernetes.io/projected/08706f9e-bd19-4f02-9d9c-27a218ec6764-kube-api-access-59tts\") pod \"08706f9e-bd19-4f02-9d9c-27a218ec6764\" (UID: \"08706f9e-bd19-4f02-9d9c-27a218ec6764\") " Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.710767 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-utilities" (OuterVolumeSpecName: "utilities") pod "08706f9e-bd19-4f02-9d9c-27a218ec6764" (UID: "08706f9e-bd19-4f02-9d9c-27a218ec6764"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.716623 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08706f9e-bd19-4f02-9d9c-27a218ec6764-kube-api-access-59tts" (OuterVolumeSpecName: "kube-api-access-59tts") pod "08706f9e-bd19-4f02-9d9c-27a218ec6764" (UID: "08706f9e-bd19-4f02-9d9c-27a218ec6764"). InnerVolumeSpecName "kube-api-access-59tts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.738170 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08706f9e-bd19-4f02-9d9c-27a218ec6764" (UID: "08706f9e-bd19-4f02-9d9c-27a218ec6764"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.813386 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59tts\" (UniqueName: \"kubernetes.io/projected/08706f9e-bd19-4f02-9d9c-27a218ec6764-kube-api-access-59tts\") on node \"crc\" DevicePath \"\"" Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.813421 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:06:14 crc kubenswrapper[4733]: I1204 20:06:14.813431 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08706f9e-bd19-4f02-9d9c-27a218ec6764-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.088765 4733 generic.go:334] "Generic (PLEG): container finished" podID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerID="58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8" exitCode=0 Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.088840 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx8z2" event={"ID":"08706f9e-bd19-4f02-9d9c-27a218ec6764","Type":"ContainerDied","Data":"58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8"} Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.088888 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx8z2" event={"ID":"08706f9e-bd19-4f02-9d9c-27a218ec6764","Type":"ContainerDied","Data":"73c9a7dcde4d1222b3692990e2af772e7265983e1f61c1c0eebb15431eebde3e"} Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.088913 4733 scope.go:117] "RemoveContainer" containerID="58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.089988 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx8z2" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.121975 4733 scope.go:117] "RemoveContainer" containerID="39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.147415 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx8z2"] Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.159088 4733 scope.go:117] "RemoveContainer" containerID="91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.159393 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx8z2"] Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.206615 4733 scope.go:117] "RemoveContainer" containerID="58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8" Dec 04 20:06:15 crc kubenswrapper[4733]: E1204 20:06:15.207258 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8\": container with ID starting with 58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8 not found: ID does not exist" containerID="58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.207321 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8"} err="failed to get container status \"58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8\": rpc error: code = NotFound desc = could not find container \"58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8\": container with ID starting with 58877d454142250f7065be8b5393d2a732ff44804861f7d0712f409462f810d8 not found: ID does not exist" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.207364 4733 scope.go:117] "RemoveContainer" containerID="39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e" Dec 04 20:06:15 crc kubenswrapper[4733]: E1204 20:06:15.207946 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e\": container with ID starting with 39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e not found: ID does not exist" containerID="39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.207984 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e"} err="failed to get container status \"39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e\": rpc error: code = NotFound desc = could not find container \"39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e\": container with ID starting with 39c7e57e9b0add917280608772ed9869666cd24d1f6236e12e68d83de527218e not found: ID does not exist" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.208015 4733 scope.go:117] "RemoveContainer" containerID="91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3" Dec 04 20:06:15 crc kubenswrapper[4733]: E1204 20:06:15.208316 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3\": container with ID starting with 91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3 not found: ID does not exist" containerID="91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3" Dec 04 20:06:15 crc kubenswrapper[4733]: I1204 20:06:15.208362 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3"} err="failed to get container status \"91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3\": rpc error: code = NotFound desc = could not find container \"91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3\": container with ID starting with 91bd1e56e95ac405bfd54c0270d67d5b59127428c5377bc40cd586ecca98cfc3 not found: ID does not exist" Dec 04 20:06:16 crc kubenswrapper[4733]: I1204 20:06:16.348333 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" path="/var/lib/kubelet/pods/08706f9e-bd19-4f02-9d9c-27a218ec6764/volumes" Dec 04 20:06:20 crc kubenswrapper[4733]: I1204 20:06:20.336303 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:06:21 crc kubenswrapper[4733]: I1204 20:06:21.169654 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"572879499a8883857ac6943ba4bc1abbd0913f8c8ec6bb99ba6a90c36d2718a8"} Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.668697 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n9fxs"] Dec 04 20:07:14 crc kubenswrapper[4733]: E1204 20:07:14.669683 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerName="extract-content" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.669698 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerName="extract-content" Dec 04 20:07:14 crc kubenswrapper[4733]: E1204 20:07:14.669740 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerName="registry-server" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.669746 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerName="registry-server" Dec 04 20:07:14 crc kubenswrapper[4733]: E1204 20:07:14.669767 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerName="extract-utilities" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.669774 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerName="extract-utilities" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.670018 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="08706f9e-bd19-4f02-9d9c-27a218ec6764" containerName="registry-server" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.671738 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.684931 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n9fxs"] Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.713842 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-utilities\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.714294 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-catalog-content\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.714462 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkk4t\" (UniqueName: \"kubernetes.io/projected/253e0a18-6ab4-4a29-a606-69354dbe6bb2-kube-api-access-zkk4t\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.815646 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-utilities\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.815703 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-catalog-content\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.815757 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkk4t\" (UniqueName: \"kubernetes.io/projected/253e0a18-6ab4-4a29-a606-69354dbe6bb2-kube-api-access-zkk4t\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.816380 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-utilities\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.816844 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-catalog-content\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:14 crc kubenswrapper[4733]: I1204 20:07:14.843462 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkk4t\" (UniqueName: \"kubernetes.io/projected/253e0a18-6ab4-4a29-a606-69354dbe6bb2-kube-api-access-zkk4t\") pod \"community-operators-n9fxs\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:15 crc kubenswrapper[4733]: I1204 20:07:15.013234 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:15 crc kubenswrapper[4733]: I1204 20:07:15.574437 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n9fxs"] Dec 04 20:07:15 crc kubenswrapper[4733]: I1204 20:07:15.814165 4733 generic.go:334] "Generic (PLEG): container finished" podID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerID="a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e" exitCode=0 Dec 04 20:07:15 crc kubenswrapper[4733]: I1204 20:07:15.814207 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9fxs" event={"ID":"253e0a18-6ab4-4a29-a606-69354dbe6bb2","Type":"ContainerDied","Data":"a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e"} Dec 04 20:07:15 crc kubenswrapper[4733]: I1204 20:07:15.814241 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9fxs" event={"ID":"253e0a18-6ab4-4a29-a606-69354dbe6bb2","Type":"ContainerStarted","Data":"96fda5ee884ba4060f0fa02bf06ec7bea3335ede7fc8cf2df16b57797e2d5c99"} Dec 04 20:07:16 crc kubenswrapper[4733]: I1204 20:07:16.830490 4733 generic.go:334] "Generic (PLEG): container finished" podID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerID="94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476" exitCode=0 Dec 04 20:07:16 crc kubenswrapper[4733]: I1204 20:07:16.830565 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9fxs" event={"ID":"253e0a18-6ab4-4a29-a606-69354dbe6bb2","Type":"ContainerDied","Data":"94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476"} Dec 04 20:07:17 crc kubenswrapper[4733]: I1204 20:07:17.856108 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9fxs" event={"ID":"253e0a18-6ab4-4a29-a606-69354dbe6bb2","Type":"ContainerStarted","Data":"75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050"} Dec 04 20:07:17 crc kubenswrapper[4733]: I1204 20:07:17.889193 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n9fxs" podStartSLOduration=2.461580041 podStartE2EDuration="3.889177872s" podCreationTimestamp="2025-12-04 20:07:14 +0000 UTC" firstStartedPulling="2025-12-04 20:07:15.816659593 +0000 UTC m=+8897.772020639" lastFinishedPulling="2025-12-04 20:07:17.244257424 +0000 UTC m=+8899.199618470" observedRunningTime="2025-12-04 20:07:17.885111051 +0000 UTC m=+8899.840472107" watchObservedRunningTime="2025-12-04 20:07:17.889177872 +0000 UTC m=+8899.844538918" Dec 04 20:07:25 crc kubenswrapper[4733]: I1204 20:07:25.013434 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:25 crc kubenswrapper[4733]: I1204 20:07:25.014038 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:25 crc kubenswrapper[4733]: I1204 20:07:25.075522 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:25 crc kubenswrapper[4733]: I1204 20:07:25.620350 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:25 crc kubenswrapper[4733]: I1204 20:07:25.670762 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n9fxs"] Dec 04 20:07:27 crc kubenswrapper[4733]: I1204 20:07:27.583409 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n9fxs" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerName="registry-server" containerID="cri-o://75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050" gracePeriod=2 Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.055251 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.232624 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkk4t\" (UniqueName: \"kubernetes.io/projected/253e0a18-6ab4-4a29-a606-69354dbe6bb2-kube-api-access-zkk4t\") pod \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.232767 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-utilities\") pod \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.232928 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-catalog-content\") pod \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\" (UID: \"253e0a18-6ab4-4a29-a606-69354dbe6bb2\") " Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.233708 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-utilities" (OuterVolumeSpecName: "utilities") pod "253e0a18-6ab4-4a29-a606-69354dbe6bb2" (UID: "253e0a18-6ab4-4a29-a606-69354dbe6bb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.239556 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/253e0a18-6ab4-4a29-a606-69354dbe6bb2-kube-api-access-zkk4t" (OuterVolumeSpecName: "kube-api-access-zkk4t") pod "253e0a18-6ab4-4a29-a606-69354dbe6bb2" (UID: "253e0a18-6ab4-4a29-a606-69354dbe6bb2"). InnerVolumeSpecName "kube-api-access-zkk4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.293911 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "253e0a18-6ab4-4a29-a606-69354dbe6bb2" (UID: "253e0a18-6ab4-4a29-a606-69354dbe6bb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.335747 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.335806 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkk4t\" (UniqueName: \"kubernetes.io/projected/253e0a18-6ab4-4a29-a606-69354dbe6bb2-kube-api-access-zkk4t\") on node \"crc\" DevicePath \"\"" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.335820 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253e0a18-6ab4-4a29-a606-69354dbe6bb2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.596030 4733 generic.go:334] "Generic (PLEG): container finished" podID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerID="75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050" exitCode=0 Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.596076 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9fxs" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.596077 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9fxs" event={"ID":"253e0a18-6ab4-4a29-a606-69354dbe6bb2","Type":"ContainerDied","Data":"75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050"} Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.596105 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9fxs" event={"ID":"253e0a18-6ab4-4a29-a606-69354dbe6bb2","Type":"ContainerDied","Data":"96fda5ee884ba4060f0fa02bf06ec7bea3335ede7fc8cf2df16b57797e2d5c99"} Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.596125 4733 scope.go:117] "RemoveContainer" containerID="75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.625293 4733 scope.go:117] "RemoveContainer" containerID="94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.625833 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n9fxs"] Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.635289 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n9fxs"] Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.657383 4733 scope.go:117] "RemoveContainer" containerID="a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.692364 4733 scope.go:117] "RemoveContainer" containerID="75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050" Dec 04 20:07:28 crc kubenswrapper[4733]: E1204 20:07:28.692989 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050\": container with ID starting with 75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050 not found: ID does not exist" containerID="75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.693022 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050"} err="failed to get container status \"75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050\": rpc error: code = NotFound desc = could not find container \"75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050\": container with ID starting with 75b36282b23d781e308c4e19298ee8738849d226cd867a8cad86d75c59a05050 not found: ID does not exist" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.693044 4733 scope.go:117] "RemoveContainer" containerID="94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476" Dec 04 20:07:28 crc kubenswrapper[4733]: E1204 20:07:28.693305 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476\": container with ID starting with 94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476 not found: ID does not exist" containerID="94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.693334 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476"} err="failed to get container status \"94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476\": rpc error: code = NotFound desc = could not find container \"94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476\": container with ID starting with 94c074d0f839917e4f2538289dd67c35d7c6f1c61523e8fa3622e85b4e80f476 not found: ID does not exist" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.693348 4733 scope.go:117] "RemoveContainer" containerID="a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e" Dec 04 20:07:28 crc kubenswrapper[4733]: E1204 20:07:28.693568 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e\": container with ID starting with a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e not found: ID does not exist" containerID="a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e" Dec 04 20:07:28 crc kubenswrapper[4733]: I1204 20:07:28.693610 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e"} err="failed to get container status \"a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e\": rpc error: code = NotFound desc = could not find container \"a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e\": container with ID starting with a3c85ca71b0d86938f386b1bbedadb593c21a2289c6f9c7df2f29a7d9e23344e not found: ID does not exist" Dec 04 20:07:30 crc kubenswrapper[4733]: I1204 20:07:30.356722 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" path="/var/lib/kubelet/pods/253e0a18-6ab4-4a29-a606-69354dbe6bb2/volumes" Dec 04 20:07:41 crc kubenswrapper[4733]: I1204 20:07:41.902612 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d72xj"] Dec 04 20:07:41 crc kubenswrapper[4733]: E1204 20:07:41.904725 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerName="extract-utilities" Dec 04 20:07:41 crc kubenswrapper[4733]: I1204 20:07:41.904762 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerName="extract-utilities" Dec 04 20:07:41 crc kubenswrapper[4733]: E1204 20:07:41.904791 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerName="registry-server" Dec 04 20:07:41 crc kubenswrapper[4733]: I1204 20:07:41.904947 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerName="registry-server" Dec 04 20:07:41 crc kubenswrapper[4733]: E1204 20:07:41.905006 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerName="extract-content" Dec 04 20:07:41 crc kubenswrapper[4733]: I1204 20:07:41.905021 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerName="extract-content" Dec 04 20:07:41 crc kubenswrapper[4733]: I1204 20:07:41.905460 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="253e0a18-6ab4-4a29-a606-69354dbe6bb2" containerName="registry-server" Dec 04 20:07:41 crc kubenswrapper[4733]: I1204 20:07:41.908770 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:41 crc kubenswrapper[4733]: I1204 20:07:41.917289 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d72xj"] Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.047048 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cnmf\" (UniqueName: \"kubernetes.io/projected/37166949-aca9-4488-9c46-818c4d8985ef-kube-api-access-7cnmf\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.047269 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-catalog-content\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.047314 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-utilities\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.149261 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cnmf\" (UniqueName: \"kubernetes.io/projected/37166949-aca9-4488-9c46-818c4d8985ef-kube-api-access-7cnmf\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.149420 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-catalog-content\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.149456 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-utilities\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.150092 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-utilities\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.150092 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-catalog-content\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.643570 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cnmf\" (UniqueName: \"kubernetes.io/projected/37166949-aca9-4488-9c46-818c4d8985ef-kube-api-access-7cnmf\") pod \"redhat-operators-d72xj\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:42 crc kubenswrapper[4733]: I1204 20:07:42.848148 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:43 crc kubenswrapper[4733]: I1204 20:07:43.155027 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d72xj"] Dec 04 20:07:43 crc kubenswrapper[4733]: I1204 20:07:43.772482 4733 generic.go:334] "Generic (PLEG): container finished" podID="37166949-aca9-4488-9c46-818c4d8985ef" containerID="fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f" exitCode=0 Dec 04 20:07:43 crc kubenswrapper[4733]: I1204 20:07:43.772594 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d72xj" event={"ID":"37166949-aca9-4488-9c46-818c4d8985ef","Type":"ContainerDied","Data":"fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f"} Dec 04 20:07:43 crc kubenswrapper[4733]: I1204 20:07:43.772908 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d72xj" event={"ID":"37166949-aca9-4488-9c46-818c4d8985ef","Type":"ContainerStarted","Data":"3a8d4b9c71d4a42dc0ae57cc06f78fc77b6db7e5b05ad9dc8883bbc3a98aee42"} Dec 04 20:07:45 crc kubenswrapper[4733]: I1204 20:07:45.797245 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d72xj" event={"ID":"37166949-aca9-4488-9c46-818c4d8985ef","Type":"ContainerStarted","Data":"0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199"} Dec 04 20:07:48 crc kubenswrapper[4733]: I1204 20:07:48.845577 4733 generic.go:334] "Generic (PLEG): container finished" podID="37166949-aca9-4488-9c46-818c4d8985ef" containerID="0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199" exitCode=0 Dec 04 20:07:48 crc kubenswrapper[4733]: I1204 20:07:48.845675 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d72xj" event={"ID":"37166949-aca9-4488-9c46-818c4d8985ef","Type":"ContainerDied","Data":"0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199"} Dec 04 20:07:49 crc kubenswrapper[4733]: I1204 20:07:49.857224 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d72xj" event={"ID":"37166949-aca9-4488-9c46-818c4d8985ef","Type":"ContainerStarted","Data":"75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac"} Dec 04 20:07:49 crc kubenswrapper[4733]: I1204 20:07:49.888545 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d72xj" podStartSLOduration=3.194456731 podStartE2EDuration="8.8885185s" podCreationTimestamp="2025-12-04 20:07:41 +0000 UTC" firstStartedPulling="2025-12-04 20:07:43.775018573 +0000 UTC m=+8925.730379619" lastFinishedPulling="2025-12-04 20:07:49.469080342 +0000 UTC m=+8931.424441388" observedRunningTime="2025-12-04 20:07:49.884981223 +0000 UTC m=+8931.840342279" watchObservedRunningTime="2025-12-04 20:07:49.8885185 +0000 UTC m=+8931.843879566" Dec 04 20:07:52 crc kubenswrapper[4733]: I1204 20:07:52.848474 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:52 crc kubenswrapper[4733]: I1204 20:07:52.849001 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:07:53 crc kubenswrapper[4733]: I1204 20:07:53.901750 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d72xj" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="registry-server" probeResult="failure" output=< Dec 04 20:07:53 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 20:07:53 crc kubenswrapper[4733]: > Dec 04 20:08:00 crc kubenswrapper[4733]: I1204 20:08:00.996089 4733 generic.go:334] "Generic (PLEG): container finished" podID="af2a17da-c85e-42d9-ab67-5c44b5c5eab8" containerID="fd9f2a61ffecf02719349b6b59db8af58812bb83d1e8adfd48ec26c02b7da01a" exitCode=0 Dec 04 20:08:00 crc kubenswrapper[4733]: I1204 20:08:00.996221 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" event={"ID":"af2a17da-c85e-42d9-ab67-5c44b5c5eab8","Type":"ContainerDied","Data":"fd9f2a61ffecf02719349b6b59db8af58812bb83d1e8adfd48ec26c02b7da01a"} Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.496711 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.617786 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ssh-key\") pod \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.618454 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhwh5\" (UniqueName: \"kubernetes.io/projected/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-kube-api-access-xhwh5\") pod \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.618609 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-agent-neutron-config-0\") pod \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.618828 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-inventory\") pod \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.618983 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ceph\") pod \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.619213 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-combined-ca-bundle\") pod \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\" (UID: \"af2a17da-c85e-42d9-ab67-5c44b5c5eab8\") " Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.624568 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ceph" (OuterVolumeSpecName: "ceph") pod "af2a17da-c85e-42d9-ab67-5c44b5c5eab8" (UID: "af2a17da-c85e-42d9-ab67-5c44b5c5eab8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.627885 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "af2a17da-c85e-42d9-ab67-5c44b5c5eab8" (UID: "af2a17da-c85e-42d9-ab67-5c44b5c5eab8"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.630079 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-kube-api-access-xhwh5" (OuterVolumeSpecName: "kube-api-access-xhwh5") pod "af2a17da-c85e-42d9-ab67-5c44b5c5eab8" (UID: "af2a17da-c85e-42d9-ab67-5c44b5c5eab8"). InnerVolumeSpecName "kube-api-access-xhwh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.648689 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-inventory" (OuterVolumeSpecName: "inventory") pod "af2a17da-c85e-42d9-ab67-5c44b5c5eab8" (UID: "af2a17da-c85e-42d9-ab67-5c44b5c5eab8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.653037 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "af2a17da-c85e-42d9-ab67-5c44b5c5eab8" (UID: "af2a17da-c85e-42d9-ab67-5c44b5c5eab8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.669398 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "af2a17da-c85e-42d9-ab67-5c44b5c5eab8" (UID: "af2a17da-c85e-42d9-ab67-5c44b5c5eab8"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.722037 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhwh5\" (UniqueName: \"kubernetes.io/projected/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-kube-api-access-xhwh5\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.722083 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.722099 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.722111 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.722125 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.722138 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af2a17da-c85e-42d9-ab67-5c44b5c5eab8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:02 crc kubenswrapper[4733]: I1204 20:08:02.979507 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.050050 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" event={"ID":"af2a17da-c85e-42d9-ab67-5c44b5c5eab8","Type":"ContainerDied","Data":"84b2d9a0e800a66fafe94424ef1002e0500a35aaaf8127210e3e6f7ca5fe2479"} Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.050090 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84b2d9a0e800a66fafe94424ef1002e0500a35aaaf8127210e3e6f7ca5fe2479" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.050165 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-jgtmh" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.110977 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.129947 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f"] Dec 04 20:08:03 crc kubenswrapper[4733]: E1204 20:08:03.130360 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2a17da-c85e-42d9-ab67-5c44b5c5eab8" containerName="neutron-sriov-openstack-openstack-cell1" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.130375 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2a17da-c85e-42d9-ab67-5c44b5c5eab8" containerName="neutron-sriov-openstack-openstack-cell1" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.130609 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="af2a17da-c85e-42d9-ab67-5c44b5c5eab8" containerName="neutron-sriov-openstack-openstack-cell1" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.131287 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.134905 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.134945 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.135015 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.139984 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.147625 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f"] Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.177984 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.233620 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.234004 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.234192 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb79q\" (UniqueName: \"kubernetes.io/projected/4bdd1faa-153d-4a64-885a-5f10370c668e-kube-api-access-lb79q\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.234301 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.234388 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.234518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.238004 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d72xj"] Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.337015 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.337082 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb79q\" (UniqueName: \"kubernetes.io/projected/4bdd1faa-153d-4a64-885a-5f10370c668e-kube-api-access-lb79q\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.337116 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.337147 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.337181 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.337221 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.341469 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.341708 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.342747 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.345076 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.346985 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.355745 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb79q\" (UniqueName: \"kubernetes.io/projected/4bdd1faa-153d-4a64-885a-5f10370c668e-kube-api-access-lb79q\") pod \"neutron-dhcp-openstack-openstack-cell1-h9s8f\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:03 crc kubenswrapper[4733]: I1204 20:08:03.493390 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.064058 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d72xj" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="registry-server" containerID="cri-o://75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac" gracePeriod=2 Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.093522 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f"] Dec 04 20:08:04 crc kubenswrapper[4733]: W1204 20:08:04.097938 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bdd1faa_153d_4a64_885a_5f10370c668e.slice/crio-53b730f6ae6d1c9c0fbc5710ae325869e11fcb4358a6e969ba6ae79180f5cd94 WatchSource:0}: Error finding container 53b730f6ae6d1c9c0fbc5710ae325869e11fcb4358a6e969ba6ae79180f5cd94: Status 404 returned error can't find the container with id 53b730f6ae6d1c9c0fbc5710ae325869e11fcb4358a6e969ba6ae79180f5cd94 Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.626190 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.664582 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-catalog-content\") pod \"37166949-aca9-4488-9c46-818c4d8985ef\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.664636 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cnmf\" (UniqueName: \"kubernetes.io/projected/37166949-aca9-4488-9c46-818c4d8985ef-kube-api-access-7cnmf\") pod \"37166949-aca9-4488-9c46-818c4d8985ef\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.664746 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-utilities\") pod \"37166949-aca9-4488-9c46-818c4d8985ef\" (UID: \"37166949-aca9-4488-9c46-818c4d8985ef\") " Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.670599 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-utilities" (OuterVolumeSpecName: "utilities") pod "37166949-aca9-4488-9c46-818c4d8985ef" (UID: "37166949-aca9-4488-9c46-818c4d8985ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.675126 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37166949-aca9-4488-9c46-818c4d8985ef-kube-api-access-7cnmf" (OuterVolumeSpecName: "kube-api-access-7cnmf") pod "37166949-aca9-4488-9c46-818c4d8985ef" (UID: "37166949-aca9-4488-9c46-818c4d8985ef"). InnerVolumeSpecName "kube-api-access-7cnmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.771279 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cnmf\" (UniqueName: \"kubernetes.io/projected/37166949-aca9-4488-9c46-818c4d8985ef-kube-api-access-7cnmf\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.771309 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.784575 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37166949-aca9-4488-9c46-818c4d8985ef" (UID: "37166949-aca9-4488-9c46-818c4d8985ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:08:04 crc kubenswrapper[4733]: I1204 20:08:04.873189 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37166949-aca9-4488-9c46-818c4d8985ef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.079695 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" event={"ID":"4bdd1faa-153d-4a64-885a-5f10370c668e","Type":"ContainerStarted","Data":"53b730f6ae6d1c9c0fbc5710ae325869e11fcb4358a6e969ba6ae79180f5cd94"} Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.083081 4733 generic.go:334] "Generic (PLEG): container finished" podID="37166949-aca9-4488-9c46-818c4d8985ef" containerID="75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac" exitCode=0 Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.083119 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d72xj" event={"ID":"37166949-aca9-4488-9c46-818c4d8985ef","Type":"ContainerDied","Data":"75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac"} Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.083142 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d72xj" event={"ID":"37166949-aca9-4488-9c46-818c4d8985ef","Type":"ContainerDied","Data":"3a8d4b9c71d4a42dc0ae57cc06f78fc77b6db7e5b05ad9dc8883bbc3a98aee42"} Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.083162 4733 scope.go:117] "RemoveContainer" containerID="75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.083314 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d72xj" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.116270 4733 scope.go:117] "RemoveContainer" containerID="0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.122169 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d72xj"] Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.132731 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d72xj"] Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.390662 4733 scope.go:117] "RemoveContainer" containerID="fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.445448 4733 scope.go:117] "RemoveContainer" containerID="75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac" Dec 04 20:08:05 crc kubenswrapper[4733]: E1204 20:08:05.445838 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac\": container with ID starting with 75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac not found: ID does not exist" containerID="75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.445878 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac"} err="failed to get container status \"75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac\": rpc error: code = NotFound desc = could not find container \"75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac\": container with ID starting with 75449046c69a13b413b6ed94bccb95b0b68dacce324f6537f49932a3eb6d6eac not found: ID does not exist" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.445903 4733 scope.go:117] "RemoveContainer" containerID="0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199" Dec 04 20:08:05 crc kubenswrapper[4733]: E1204 20:08:05.446404 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199\": container with ID starting with 0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199 not found: ID does not exist" containerID="0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.446439 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199"} err="failed to get container status \"0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199\": rpc error: code = NotFound desc = could not find container \"0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199\": container with ID starting with 0e7266da2223e94b46ded6810eca2846206911c443ad35b3442407f653856199 not found: ID does not exist" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.446466 4733 scope.go:117] "RemoveContainer" containerID="fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f" Dec 04 20:08:05 crc kubenswrapper[4733]: E1204 20:08:05.447071 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f\": container with ID starting with fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f not found: ID does not exist" containerID="fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f" Dec 04 20:08:05 crc kubenswrapper[4733]: I1204 20:08:05.447108 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f"} err="failed to get container status \"fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f\": rpc error: code = NotFound desc = could not find container \"fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f\": container with ID starting with fd52c625b980728ce655f69d75053ed4fe68f0707ead0f8ca858013026352c8f not found: ID does not exist" Dec 04 20:08:06 crc kubenswrapper[4733]: I1204 20:08:06.108317 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" event={"ID":"4bdd1faa-153d-4a64-885a-5f10370c668e","Type":"ContainerStarted","Data":"ac494f95d4b1f5d50353e8df3687f28e4a2fc846aad86f4c262a8ab0d2b31cf0"} Dec 04 20:08:06 crc kubenswrapper[4733]: I1204 20:08:06.130787 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" podStartSLOduration=2.624296154 podStartE2EDuration="3.130767111s" podCreationTimestamp="2025-12-04 20:08:03 +0000 UTC" firstStartedPulling="2025-12-04 20:08:04.106067949 +0000 UTC m=+8946.061428995" lastFinishedPulling="2025-12-04 20:08:04.612538916 +0000 UTC m=+8946.567899952" observedRunningTime="2025-12-04 20:08:06.129265341 +0000 UTC m=+8948.084626397" watchObservedRunningTime="2025-12-04 20:08:06.130767111 +0000 UTC m=+8948.086128157" Dec 04 20:08:06 crc kubenswrapper[4733]: I1204 20:08:06.347562 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37166949-aca9-4488-9c46-818c4d8985ef" path="/var/lib/kubelet/pods/37166949-aca9-4488-9c46-818c4d8985ef/volumes" Dec 04 20:08:45 crc kubenswrapper[4733]: I1204 20:08:45.362253 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:08:45 crc kubenswrapper[4733]: I1204 20:08:45.362976 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:09:15 crc kubenswrapper[4733]: I1204 20:09:15.362101 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:09:15 crc kubenswrapper[4733]: I1204 20:09:15.362716 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:09:29 crc kubenswrapper[4733]: I1204 20:09:29.268587 4733 scope.go:117] "RemoveContainer" containerID="efa4771cc4824329132444d78e03abbb67d6577ef7b3a69256616bde3dedae06" Dec 04 20:09:29 crc kubenswrapper[4733]: I1204 20:09:29.456992 4733 scope.go:117] "RemoveContainer" containerID="a8bbf759b1b623af21418f0018e669ee2209b9a805107c6e81c128848da38ccd" Dec 04 20:09:29 crc kubenswrapper[4733]: I1204 20:09:29.520242 4733 scope.go:117] "RemoveContainer" containerID="5b45936599b2a51ceaad1637192107cdb6d8abc7970068d5a905f493dbd1b783" Dec 04 20:09:45 crc kubenswrapper[4733]: I1204 20:09:45.361581 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:09:45 crc kubenswrapper[4733]: I1204 20:09:45.362151 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:09:45 crc kubenswrapper[4733]: I1204 20:09:45.362191 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 20:09:45 crc kubenswrapper[4733]: I1204 20:09:45.363001 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"572879499a8883857ac6943ba4bc1abbd0913f8c8ec6bb99ba6a90c36d2718a8"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 20:09:45 crc kubenswrapper[4733]: I1204 20:09:45.363059 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://572879499a8883857ac6943ba4bc1abbd0913f8c8ec6bb99ba6a90c36d2718a8" gracePeriod=600 Dec 04 20:09:46 crc kubenswrapper[4733]: I1204 20:09:46.237339 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="572879499a8883857ac6943ba4bc1abbd0913f8c8ec6bb99ba6a90c36d2718a8" exitCode=0 Dec 04 20:09:46 crc kubenswrapper[4733]: I1204 20:09:46.237432 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"572879499a8883857ac6943ba4bc1abbd0913f8c8ec6bb99ba6a90c36d2718a8"} Dec 04 20:09:46 crc kubenswrapper[4733]: I1204 20:09:46.237877 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039"} Dec 04 20:09:46 crc kubenswrapper[4733]: I1204 20:09:46.237903 4733 scope.go:117] "RemoveContainer" containerID="3befb51e69f711c775abe81fe49da342bcbe606288de7d5c9c56f1ac26737f1c" Dec 04 20:11:45 crc kubenswrapper[4733]: I1204 20:11:45.362451 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:11:45 crc kubenswrapper[4733]: I1204 20:11:45.363236 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:12:15 crc kubenswrapper[4733]: I1204 20:12:15.361755 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:12:15 crc kubenswrapper[4733]: I1204 20:12:15.362441 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.361670 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.362233 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.362276 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.363068 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.363122 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" gracePeriod=600 Dec 04 20:12:45 crc kubenswrapper[4733]: E1204 20:12:45.480754 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.929835 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" exitCode=0 Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.929890 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039"} Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.929932 4733 scope.go:117] "RemoveContainer" containerID="572879499a8883857ac6943ba4bc1abbd0913f8c8ec6bb99ba6a90c36d2718a8" Dec 04 20:12:45 crc kubenswrapper[4733]: I1204 20:12:45.930954 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:12:45 crc kubenswrapper[4733]: E1204 20:12:45.931629 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:12:59 crc kubenswrapper[4733]: I1204 20:12:59.336122 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:12:59 crc kubenswrapper[4733]: E1204 20:12:59.337219 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:13:05 crc kubenswrapper[4733]: I1204 20:13:05.154888 4733 generic.go:334] "Generic (PLEG): container finished" podID="4bdd1faa-153d-4a64-885a-5f10370c668e" containerID="ac494f95d4b1f5d50353e8df3687f28e4a2fc846aad86f4c262a8ab0d2b31cf0" exitCode=0 Dec 04 20:13:05 crc kubenswrapper[4733]: I1204 20:13:05.154977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" event={"ID":"4bdd1faa-153d-4a64-885a-5f10370c668e","Type":"ContainerDied","Data":"ac494f95d4b1f5d50353e8df3687f28e4a2fc846aad86f4c262a8ab0d2b31cf0"} Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.691530 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.816812 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-agent-neutron-config-0\") pod \"4bdd1faa-153d-4a64-885a-5f10370c668e\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.816987 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-combined-ca-bundle\") pod \"4bdd1faa-153d-4a64-885a-5f10370c668e\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.817045 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ceph\") pod \"4bdd1faa-153d-4a64-885a-5f10370c668e\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.817101 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ssh-key\") pod \"4bdd1faa-153d-4a64-885a-5f10370c668e\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.817269 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb79q\" (UniqueName: \"kubernetes.io/projected/4bdd1faa-153d-4a64-885a-5f10370c668e-kube-api-access-lb79q\") pod \"4bdd1faa-153d-4a64-885a-5f10370c668e\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.817308 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-inventory\") pod \"4bdd1faa-153d-4a64-885a-5f10370c668e\" (UID: \"4bdd1faa-153d-4a64-885a-5f10370c668e\") " Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.824183 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ceph" (OuterVolumeSpecName: "ceph") pod "4bdd1faa-153d-4a64-885a-5f10370c668e" (UID: "4bdd1faa-153d-4a64-885a-5f10370c668e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.824326 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "4bdd1faa-153d-4a64-885a-5f10370c668e" (UID: "4bdd1faa-153d-4a64-885a-5f10370c668e"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.827029 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bdd1faa-153d-4a64-885a-5f10370c668e-kube-api-access-lb79q" (OuterVolumeSpecName: "kube-api-access-lb79q") pod "4bdd1faa-153d-4a64-885a-5f10370c668e" (UID: "4bdd1faa-153d-4a64-885a-5f10370c668e"). InnerVolumeSpecName "kube-api-access-lb79q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.851008 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "4bdd1faa-153d-4a64-885a-5f10370c668e" (UID: "4bdd1faa-153d-4a64-885a-5f10370c668e"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.853939 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-inventory" (OuterVolumeSpecName: "inventory") pod "4bdd1faa-153d-4a64-885a-5f10370c668e" (UID: "4bdd1faa-153d-4a64-885a-5f10370c668e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.859921 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4bdd1faa-153d-4a64-885a-5f10370c668e" (UID: "4bdd1faa-153d-4a64-885a-5f10370c668e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.920560 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.920608 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.921701 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.921728 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb79q\" (UniqueName: \"kubernetes.io/projected/4bdd1faa-153d-4a64-885a-5f10370c668e-kube-api-access-lb79q\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.921740 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:06 crc kubenswrapper[4733]: I1204 20:13:06.921751 4733 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4bdd1faa-153d-4a64-885a-5f10370c668e-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:07 crc kubenswrapper[4733]: I1204 20:13:07.181122 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" event={"ID":"4bdd1faa-153d-4a64-885a-5f10370c668e","Type":"ContainerDied","Data":"53b730f6ae6d1c9c0fbc5710ae325869e11fcb4358a6e969ba6ae79180f5cd94"} Dec 04 20:13:07 crc kubenswrapper[4733]: I1204 20:13:07.181187 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53b730f6ae6d1c9c0fbc5710ae325869e11fcb4358a6e969ba6ae79180f5cd94" Dec 04 20:13:07 crc kubenswrapper[4733]: I1204 20:13:07.181194 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-h9s8f" Dec 04 20:13:10 crc kubenswrapper[4733]: I1204 20:13:10.001184 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 20:13:10 crc kubenswrapper[4733]: I1204 20:13:10.001931 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="ef61747b-f475-4934-974d-7cf53750963b" containerName="nova-cell0-conductor-conductor" containerID="cri-o://f47f2b93431904bee107f24ffa8f58bf14b54ed1f3f0ce1b95d4d0d724abc65d" gracePeriod=30 Dec 04 20:13:10 crc kubenswrapper[4733]: I1204 20:13:10.017881 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 20:13:10 crc kubenswrapper[4733]: I1204 20:13:10.018258 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="c2d32ba0-df74-456e-a81b-784712deb2eb" containerName="nova-cell1-conductor-conductor" containerID="cri-o://019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee" gracePeriod=30 Dec 04 20:13:10 crc kubenswrapper[4733]: I1204 20:13:10.342115 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:13:10 crc kubenswrapper[4733]: E1204 20:13:10.342618 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:13:10 crc kubenswrapper[4733]: E1204 20:13:10.435571 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f47f2b93431904bee107f24ffa8f58bf14b54ed1f3f0ce1b95d4d0d724abc65d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 20:13:10 crc kubenswrapper[4733]: E1204 20:13:10.437075 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f47f2b93431904bee107f24ffa8f58bf14b54ed1f3f0ce1b95d4d0d724abc65d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 20:13:10 crc kubenswrapper[4733]: E1204 20:13:10.438608 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f47f2b93431904bee107f24ffa8f58bf14b54ed1f3f0ce1b95d4d0d724abc65d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 20:13:10 crc kubenswrapper[4733]: E1204 20:13:10.438673 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="ef61747b-f475-4934-974d-7cf53750963b" containerName="nova-cell0-conductor-conductor" Dec 04 20:13:11 crc kubenswrapper[4733]: I1204 20:13:11.159362 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 20:13:11 crc kubenswrapper[4733]: I1204 20:13:11.159869 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-log" containerID="cri-o://6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149" gracePeriod=30 Dec 04 20:13:11 crc kubenswrapper[4733]: I1204 20:13:11.159984 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-api" containerID="cri-o://e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e" gracePeriod=30 Dec 04 20:13:11 crc kubenswrapper[4733]: I1204 20:13:11.198593 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 20:13:11 crc kubenswrapper[4733]: I1204 20:13:11.206859 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="fffc6200-ffec-4120-9e25-feb4982180c3" containerName="nova-scheduler-scheduler" containerID="cri-o://0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2" gracePeriod=30 Dec 04 20:13:11 crc kubenswrapper[4733]: I1204 20:13:11.260437 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 20:13:11 crc kubenswrapper[4733]: I1204 20:13:11.260699 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-log" containerID="cri-o://6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205" gracePeriod=30 Dec 04 20:13:11 crc kubenswrapper[4733]: I1204 20:13:11.260882 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-metadata" containerID="cri-o://5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876" gracePeriod=30 Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.129157 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.238553 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-combined-ca-bundle\") pod \"c2d32ba0-df74-456e-a81b-784712deb2eb\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.238993 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-config-data\") pod \"c2d32ba0-df74-456e-a81b-784712deb2eb\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.239088 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g76fr\" (UniqueName: \"kubernetes.io/projected/c2d32ba0-df74-456e-a81b-784712deb2eb-kube-api-access-g76fr\") pod \"c2d32ba0-df74-456e-a81b-784712deb2eb\" (UID: \"c2d32ba0-df74-456e-a81b-784712deb2eb\") " Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.245033 4733 generic.go:334] "Generic (PLEG): container finished" podID="b248d65a-3d4e-4c73-a49c-53864042c710" containerID="6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205" exitCode=143 Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.245123 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b248d65a-3d4e-4c73-a49c-53864042c710","Type":"ContainerDied","Data":"6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205"} Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.245224 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2d32ba0-df74-456e-a81b-784712deb2eb-kube-api-access-g76fr" (OuterVolumeSpecName: "kube-api-access-g76fr") pod "c2d32ba0-df74-456e-a81b-784712deb2eb" (UID: "c2d32ba0-df74-456e-a81b-784712deb2eb"). InnerVolumeSpecName "kube-api-access-g76fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.248090 4733 generic.go:334] "Generic (PLEG): container finished" podID="c2d32ba0-df74-456e-a81b-784712deb2eb" containerID="019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee" exitCode=0 Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.248149 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c2d32ba0-df74-456e-a81b-784712deb2eb","Type":"ContainerDied","Data":"019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee"} Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.248172 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c2d32ba0-df74-456e-a81b-784712deb2eb","Type":"ContainerDied","Data":"2b5c971cc5e33025ca24836b121c7295ae7509a03e1447c708bb3ba30276def4"} Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.248192 4733 scope.go:117] "RemoveContainer" containerID="019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.248254 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.250676 4733 generic.go:334] "Generic (PLEG): container finished" podID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerID="6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149" exitCode=143 Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.250713 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60","Type":"ContainerDied","Data":"6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149"} Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.271542 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-config-data" (OuterVolumeSpecName: "config-data") pod "c2d32ba0-df74-456e-a81b-784712deb2eb" (UID: "c2d32ba0-df74-456e-a81b-784712deb2eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.278141 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2d32ba0-df74-456e-a81b-784712deb2eb" (UID: "c2d32ba0-df74-456e-a81b-784712deb2eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.280316 4733 scope.go:117] "RemoveContainer" containerID="019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee" Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.280925 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee\": container with ID starting with 019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee not found: ID does not exist" containerID="019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.280968 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee"} err="failed to get container status \"019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee\": rpc error: code = NotFound desc = could not find container \"019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee\": container with ID starting with 019f9585ec2ad97df5022c3750d32e4d848284b857fad2871d5e5061f928ccee not found: ID does not exist" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.342132 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.342460 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d32ba0-df74-456e-a81b-784712deb2eb-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.342475 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g76fr\" (UniqueName: \"kubernetes.io/projected/c2d32ba0-df74-456e-a81b-784712deb2eb-kube-api-access-g76fr\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.572423 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.574363 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.574734 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.578053 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.578127 4733 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="fffc6200-ffec-4120-9e25-feb4982180c3" containerName="nova-scheduler-scheduler" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.586198 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.603670 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.604187 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="extract-content" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.604211 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="extract-content" Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.604226 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdd1faa-153d-4a64-885a-5f10370c668e" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.604235 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdd1faa-153d-4a64-885a-5f10370c668e" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.604264 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="extract-utilities" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.604272 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="extract-utilities" Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.604304 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d32ba0-df74-456e-a81b-784712deb2eb" containerName="nova-cell1-conductor-conductor" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.604311 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d32ba0-df74-456e-a81b-784712deb2eb" containerName="nova-cell1-conductor-conductor" Dec 04 20:13:12 crc kubenswrapper[4733]: E1204 20:13:12.604323 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="registry-server" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.604331 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="registry-server" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.604607 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bdd1faa-153d-4a64-885a-5f10370c668e" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.604630 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="37166949-aca9-4488-9c46-818c4d8985ef" containerName="registry-server" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.604655 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d32ba0-df74-456e-a81b-784712deb2eb" containerName="nova-cell1-conductor-conductor" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.605667 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.608591 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.624184 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.649665 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef56127d-af29-4370-9c9d-c2b72b040fac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.649974 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mmxj\" (UniqueName: \"kubernetes.io/projected/ef56127d-af29-4370-9c9d-c2b72b040fac-kube-api-access-4mmxj\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.650026 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef56127d-af29-4370-9c9d-c2b72b040fac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.751781 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mmxj\" (UniqueName: \"kubernetes.io/projected/ef56127d-af29-4370-9c9d-c2b72b040fac-kube-api-access-4mmxj\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.751869 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef56127d-af29-4370-9c9d-c2b72b040fac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.751927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef56127d-af29-4370-9c9d-c2b72b040fac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.756016 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef56127d-af29-4370-9c9d-c2b72b040fac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.756535 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef56127d-af29-4370-9c9d-c2b72b040fac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.781008 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mmxj\" (UniqueName: \"kubernetes.io/projected/ef56127d-af29-4370-9c9d-c2b72b040fac-kube-api-access-4mmxj\") pod \"nova-cell1-conductor-0\" (UID: \"ef56127d-af29-4370-9c9d-c2b72b040fac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:12 crc kubenswrapper[4733]: I1204 20:13:12.935373 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:13 crc kubenswrapper[4733]: I1204 20:13:13.417771 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 20:13:13 crc kubenswrapper[4733]: W1204 20:13:13.418057 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef56127d_af29_4370_9c9d_c2b72b040fac.slice/crio-ef4f16c9668eda3487a298cb78526efcd2cf79437996e388a49bbd924e264a77 WatchSource:0}: Error finding container ef4f16c9668eda3487a298cb78526efcd2cf79437996e388a49bbd924e264a77: Status 404 returned error can't find the container with id ef4f16c9668eda3487a298cb78526efcd2cf79437996e388a49bbd924e264a77 Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.280264 4733 generic.go:334] "Generic (PLEG): container finished" podID="ef61747b-f475-4934-974d-7cf53750963b" containerID="f47f2b93431904bee107f24ffa8f58bf14b54ed1f3f0ce1b95d4d0d724abc65d" exitCode=0 Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.280345 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ef61747b-f475-4934-974d-7cf53750963b","Type":"ContainerDied","Data":"f47f2b93431904bee107f24ffa8f58bf14b54ed1f3f0ce1b95d4d0d724abc65d"} Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.283074 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ef56127d-af29-4370-9c9d-c2b72b040fac","Type":"ContainerStarted","Data":"535c1b1b4e2bf3a30e5057120d1197900265f691b5786fcd1085c58b7e0984a3"} Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.283137 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ef56127d-af29-4370-9c9d-c2b72b040fac","Type":"ContainerStarted","Data":"ef4f16c9668eda3487a298cb78526efcd2cf79437996e388a49bbd924e264a77"} Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.283184 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.312205 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.31218577 podStartE2EDuration="2.31218577s" podCreationTimestamp="2025-12-04 20:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 20:13:14.301930061 +0000 UTC m=+9256.257291117" watchObservedRunningTime="2025-12-04 20:13:14.31218577 +0000 UTC m=+9256.267546826" Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.348346 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2d32ba0-df74-456e-a81b-784712deb2eb" path="/var/lib/kubelet/pods/c2d32ba0-df74-456e-a81b-784712deb2eb/volumes" Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.536200 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.91:8775/\": read tcp 10.217.0.2:47740->10.217.1.91:8775: read: connection reset by peer" Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.536252 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.91:8775/\": read tcp 10.217.0.2:47752->10.217.1.91:8775: read: connection reset by peer" Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.864342 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.932113 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xljmc\" (UniqueName: \"kubernetes.io/projected/ef61747b-f475-4934-974d-7cf53750963b-kube-api-access-xljmc\") pod \"ef61747b-f475-4934-974d-7cf53750963b\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.932221 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-config-data\") pod \"ef61747b-f475-4934-974d-7cf53750963b\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.932373 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-combined-ca-bundle\") pod \"ef61747b-f475-4934-974d-7cf53750963b\" (UID: \"ef61747b-f475-4934-974d-7cf53750963b\") " Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.951084 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef61747b-f475-4934-974d-7cf53750963b-kube-api-access-xljmc" (OuterVolumeSpecName: "kube-api-access-xljmc") pod "ef61747b-f475-4934-974d-7cf53750963b" (UID: "ef61747b-f475-4934-974d-7cf53750963b"). InnerVolumeSpecName "kube-api-access-xljmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.977324 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef61747b-f475-4934-974d-7cf53750963b" (UID: "ef61747b-f475-4934-974d-7cf53750963b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:14 crc kubenswrapper[4733]: I1204 20:13:14.998034 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-config-data" (OuterVolumeSpecName: "config-data") pod "ef61747b-f475-4934-974d-7cf53750963b" (UID: "ef61747b-f475-4934-974d-7cf53750963b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.028831 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.035367 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xljmc\" (UniqueName: \"kubernetes.io/projected/ef61747b-f475-4934-974d-7cf53750963b-kube-api-access-xljmc\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.035421 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.035433 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef61747b-f475-4934-974d-7cf53750963b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.136661 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b248d65a-3d4e-4c73-a49c-53864042c710-logs\") pod \"b248d65a-3d4e-4c73-a49c-53864042c710\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.136721 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-config-data\") pod \"b248d65a-3d4e-4c73-a49c-53864042c710\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.136816 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2955\" (UniqueName: \"kubernetes.io/projected/b248d65a-3d4e-4c73-a49c-53864042c710-kube-api-access-r2955\") pod \"b248d65a-3d4e-4c73-a49c-53864042c710\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.136867 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-combined-ca-bundle\") pod \"b248d65a-3d4e-4c73-a49c-53864042c710\" (UID: \"b248d65a-3d4e-4c73-a49c-53864042c710\") " Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.137845 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b248d65a-3d4e-4c73-a49c-53864042c710-logs" (OuterVolumeSpecName: "logs") pod "b248d65a-3d4e-4c73-a49c-53864042c710" (UID: "b248d65a-3d4e-4c73-a49c-53864042c710"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.140344 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.141279 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b248d65a-3d4e-4c73-a49c-53864042c710-kube-api-access-r2955" (OuterVolumeSpecName: "kube-api-access-r2955") pod "b248d65a-3d4e-4c73-a49c-53864042c710" (UID: "b248d65a-3d4e-4c73-a49c-53864042c710"). InnerVolumeSpecName "kube-api-access-r2955". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.168680 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b248d65a-3d4e-4c73-a49c-53864042c710" (UID: "b248d65a-3d4e-4c73-a49c-53864042c710"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.170840 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-config-data" (OuterVolumeSpecName: "config-data") pod "b248d65a-3d4e-4c73-a49c-53864042c710" (UID: "b248d65a-3d4e-4c73-a49c-53864042c710"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.239254 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-444dk\" (UniqueName: \"kubernetes.io/projected/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-kube-api-access-444dk\") pod \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.239709 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-logs\") pod \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.239818 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-combined-ca-bundle\") pod \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.239879 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-config-data\") pod \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\" (UID: \"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60\") " Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.240297 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-logs" (OuterVolumeSpecName: "logs") pod "d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" (UID: "d9583123-6b5c-42b8-bd1e-5a8bc63b7f60"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.242393 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-kube-api-access-444dk" (OuterVolumeSpecName: "kube-api-access-444dk") pod "d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" (UID: "d9583123-6b5c-42b8-bd1e-5a8bc63b7f60"). InnerVolumeSpecName "kube-api-access-444dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.243070 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-444dk\" (UniqueName: \"kubernetes.io/projected/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-kube-api-access-444dk\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.243099 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b248d65a-3d4e-4c73-a49c-53864042c710-logs\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.243112 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.243125 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2955\" (UniqueName: \"kubernetes.io/projected/b248d65a-3d4e-4c73-a49c-53864042c710-kube-api-access-r2955\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.243136 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-logs\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.243143 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b248d65a-3d4e-4c73-a49c-53864042c710-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.264723 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" (UID: "d9583123-6b5c-42b8-bd1e-5a8bc63b7f60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.265257 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-config-data" (OuterVolumeSpecName: "config-data") pod "d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" (UID: "d9583123-6b5c-42b8-bd1e-5a8bc63b7f60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.300971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ef61747b-f475-4934-974d-7cf53750963b","Type":"ContainerDied","Data":"6210138a8804afba410828ee650a6b740ac3fe0971496c2e8f3c23d07fbe3ae5"} Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.301019 4733 scope.go:117] "RemoveContainer" containerID="f47f2b93431904bee107f24ffa8f58bf14b54ed1f3f0ce1b95d4d0d724abc65d" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.301016 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.304350 4733 generic.go:334] "Generic (PLEG): container finished" podID="b248d65a-3d4e-4c73-a49c-53864042c710" containerID="5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876" exitCode=0 Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.304430 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b248d65a-3d4e-4c73-a49c-53864042c710","Type":"ContainerDied","Data":"5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876"} Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.304463 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b248d65a-3d4e-4c73-a49c-53864042c710","Type":"ContainerDied","Data":"37bfcf2b5a50ae75b10e7a992aea2892d9a04ba6b3c2b332de157eb88247f666"} Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.304533 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.313609 4733 generic.go:334] "Generic (PLEG): container finished" podID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerID="e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e" exitCode=0 Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.314302 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.314384 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60","Type":"ContainerDied","Data":"e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e"} Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.314414 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9583123-6b5c-42b8-bd1e-5a8bc63b7f60","Type":"ContainerDied","Data":"f663f4c4e891851f9cbead8ca9b169acdd8974b27e7b74ead0ca61c40fa952bb"} Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.328001 4733 scope.go:117] "RemoveContainer" containerID="5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.345720 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.345776 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.361134 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.361671 4733 scope.go:117] "RemoveContainer" containerID="6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.387171 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.398718 4733 scope.go:117] "RemoveContainer" containerID="5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876" Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.402278 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876\": container with ID starting with 5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876 not found: ID does not exist" containerID="5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.402331 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876"} err="failed to get container status \"5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876\": rpc error: code = NotFound desc = could not find container \"5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876\": container with ID starting with 5dd982c7fa4e9b3b777bf6027d77500710c0c2b4303aecd5d3fa3d3ba2d36876 not found: ID does not exist" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.402365 4733 scope.go:117] "RemoveContainer" containerID="6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205" Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.407979 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205\": container with ID starting with 6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205 not found: ID does not exist" containerID="6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.408038 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205"} err="failed to get container status \"6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205\": rpc error: code = NotFound desc = could not find container \"6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205\": container with ID starting with 6514d3fc3abeb262fc422f62c8169e45dc9ddabef1677689b33f237d33097205 not found: ID does not exist" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.408070 4733 scope.go:117] "RemoveContainer" containerID="e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.418871 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.470976 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.471475 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-log" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471489 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-log" Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.471506 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-metadata" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471512 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-metadata" Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.471572 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-api" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471578 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-api" Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.471592 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-log" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471597 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-log" Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.471621 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef61747b-f475-4934-974d-7cf53750963b" containerName="nova-cell0-conductor-conductor" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471627 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef61747b-f475-4934-974d-7cf53750963b" containerName="nova-cell0-conductor-conductor" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471880 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-log" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471900 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-log" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471917 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" containerName="nova-metadata-metadata" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471927 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef61747b-f475-4934-974d-7cf53750963b" containerName="nova-cell0-conductor-conductor" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.471938 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" containerName="nova-api-api" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.472813 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.474721 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.480738 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.488776 4733 scope.go:117] "RemoveContainer" containerID="6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.492708 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.502816 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.512180 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.514457 4733 scope.go:117] "RemoveContainer" containerID="e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e" Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.514952 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e\": container with ID starting with e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e not found: ID does not exist" containerID="e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.515006 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e"} err="failed to get container status \"e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e\": rpc error: code = NotFound desc = could not find container \"e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e\": container with ID starting with e094c15c0f9bf6cab622bd5f1b9c78531cbfc7f117899ac465700c719535626e not found: ID does not exist" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.515041 4733 scope.go:117] "RemoveContainer" containerID="6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149" Dec 04 20:13:15 crc kubenswrapper[4733]: E1204 20:13:15.515345 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149\": container with ID starting with 6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149 not found: ID does not exist" containerID="6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.515383 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149"} err="failed to get container status \"6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149\": rpc error: code = NotFound desc = could not find container \"6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149\": container with ID starting with 6d2dc550fcf5c7c95884c4f0c033fc906c70b0ddb132a0926bd209b4a1ed3149 not found: ID does not exist" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.523391 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.525441 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.527992 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.534614 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.547521 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.549642 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.550028 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb67p\" (UniqueName: \"kubernetes.io/projected/844a3369-74af-4519-919b-f55b202c9e5e-kube-api-access-nb67p\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.550182 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/844a3369-74af-4519-919b-f55b202c9e5e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.550215 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844a3369-74af-4519-919b-f55b202c9e5e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.551726 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.569508 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.651917 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/844a3369-74af-4519-919b-f55b202c9e5e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652284 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-config-data\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652311 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844a3369-74af-4519-919b-f55b202c9e5e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652362 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d50035b9-6d8a-4760-8088-2a9771cb63d1-logs\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652429 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d50035b9-6d8a-4760-8088-2a9771cb63d1-config-data\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652457 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-logs\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652492 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb67p\" (UniqueName: \"kubernetes.io/projected/844a3369-74af-4519-919b-f55b202c9e5e-kube-api-access-nb67p\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652580 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652637 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mlvf\" (UniqueName: \"kubernetes.io/projected/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-kube-api-access-9mlvf\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652659 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lwd4\" (UniqueName: \"kubernetes.io/projected/d50035b9-6d8a-4760-8088-2a9771cb63d1-kube-api-access-2lwd4\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.652681 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d50035b9-6d8a-4760-8088-2a9771cb63d1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.660529 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844a3369-74af-4519-919b-f55b202c9e5e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.676466 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb67p\" (UniqueName: \"kubernetes.io/projected/844a3369-74af-4519-919b-f55b202c9e5e-kube-api-access-nb67p\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.676816 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/844a3369-74af-4519-919b-f55b202c9e5e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"844a3369-74af-4519-919b-f55b202c9e5e\") " pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.753969 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-config-data\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.754037 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d50035b9-6d8a-4760-8088-2a9771cb63d1-logs\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.754076 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d50035b9-6d8a-4760-8088-2a9771cb63d1-config-data\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.754097 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-logs\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.754150 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.754190 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mlvf\" (UniqueName: \"kubernetes.io/projected/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-kube-api-access-9mlvf\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.754242 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwd4\" (UniqueName: \"kubernetes.io/projected/d50035b9-6d8a-4760-8088-2a9771cb63d1-kube-api-access-2lwd4\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.754261 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d50035b9-6d8a-4760-8088-2a9771cb63d1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.754889 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d50035b9-6d8a-4760-8088-2a9771cb63d1-logs\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.755462 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-logs\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.759475 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d50035b9-6d8a-4760-8088-2a9771cb63d1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.759478 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.769640 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d50035b9-6d8a-4760-8088-2a9771cb63d1-config-data\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.774080 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-config-data\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.784602 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mlvf\" (UniqueName: \"kubernetes.io/projected/04fab59a-c08a-4fd3-926a-51a6bd6e6d3f-kube-api-access-9mlvf\") pod \"nova-api-0\" (UID: \"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f\") " pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.786345 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwd4\" (UniqueName: \"kubernetes.io/projected/d50035b9-6d8a-4760-8088-2a9771cb63d1-kube-api-access-2lwd4\") pod \"nova-metadata-0\" (UID: \"d50035b9-6d8a-4760-8088-2a9771cb63d1\") " pod="openstack/nova-metadata-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.792842 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.925531 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 20:13:15 crc kubenswrapper[4733]: I1204 20:13:15.939443 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.348025 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b248d65a-3d4e-4c73-a49c-53864042c710" path="/var/lib/kubelet/pods/b248d65a-3d4e-4c73-a49c-53864042c710/volumes" Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.349773 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9583123-6b5c-42b8-bd1e-5a8bc63b7f60" path="/var/lib/kubelet/pods/d9583123-6b5c-42b8-bd1e-5a8bc63b7f60/volumes" Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.350569 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef61747b-f475-4934-974d-7cf53750963b" path="/var/lib/kubelet/pods/ef61747b-f475-4934-974d-7cf53750963b/volumes" Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.442777 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 20:13:16 crc kubenswrapper[4733]: W1204 20:13:16.464597 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod844a3369_74af_4519_919b_f55b202c9e5e.slice/crio-8545d4237ec997a95346870762acce5a35532a0118a2c311eafc4c948674f1f1 WatchSource:0}: Error finding container 8545d4237ec997a95346870762acce5a35532a0118a2c311eafc4c948674f1f1: Status 404 returned error can't find the container with id 8545d4237ec997a95346870762acce5a35532a0118a2c311eafc4c948674f1f1 Dec 04 20:13:16 crc kubenswrapper[4733]: W1204 20:13:16.597414 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04fab59a_c08a_4fd3_926a_51a6bd6e6d3f.slice/crio-d59bbeb71175ab2af63df7fe5b33a9fee982a34abe3fc5de9d205935398b3e0d WatchSource:0}: Error finding container d59bbeb71175ab2af63df7fe5b33a9fee982a34abe3fc5de9d205935398b3e0d: Status 404 returned error can't find the container with id d59bbeb71175ab2af63df7fe5b33a9fee982a34abe3fc5de9d205935398b3e0d Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.617108 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.625194 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.858096 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.989386 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-combined-ca-bundle\") pod \"fffc6200-ffec-4120-9e25-feb4982180c3\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.989430 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wsvh\" (UniqueName: \"kubernetes.io/projected/fffc6200-ffec-4120-9e25-feb4982180c3-kube-api-access-4wsvh\") pod \"fffc6200-ffec-4120-9e25-feb4982180c3\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.989498 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-config-data\") pod \"fffc6200-ffec-4120-9e25-feb4982180c3\" (UID: \"fffc6200-ffec-4120-9e25-feb4982180c3\") " Dec 04 20:13:16 crc kubenswrapper[4733]: I1204 20:13:16.993869 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fffc6200-ffec-4120-9e25-feb4982180c3-kube-api-access-4wsvh" (OuterVolumeSpecName: "kube-api-access-4wsvh") pod "fffc6200-ffec-4120-9e25-feb4982180c3" (UID: "fffc6200-ffec-4120-9e25-feb4982180c3"). InnerVolumeSpecName "kube-api-access-4wsvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.022292 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-config-data" (OuterVolumeSpecName: "config-data") pod "fffc6200-ffec-4120-9e25-feb4982180c3" (UID: "fffc6200-ffec-4120-9e25-feb4982180c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.023541 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fffc6200-ffec-4120-9e25-feb4982180c3" (UID: "fffc6200-ffec-4120-9e25-feb4982180c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.095163 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.095486 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffc6200-ffec-4120-9e25-feb4982180c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.095552 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wsvh\" (UniqueName: \"kubernetes.io/projected/fffc6200-ffec-4120-9e25-feb4982180c3-kube-api-access-4wsvh\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.347107 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f","Type":"ContainerStarted","Data":"48597a17b0994981b20f8a604f92222bc9389deacf4f202b88602cdcfd0b7b42"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.347153 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f","Type":"ContainerStarted","Data":"7370f1c70f13dcbda239d2e2bb60b74cddf861f56376e7d091c847cf42826490"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.347164 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"04fab59a-c08a-4fd3-926a-51a6bd6e6d3f","Type":"ContainerStarted","Data":"d59bbeb71175ab2af63df7fe5b33a9fee982a34abe3fc5de9d205935398b3e0d"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.353456 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d50035b9-6d8a-4760-8088-2a9771cb63d1","Type":"ContainerStarted","Data":"dcdf06f57618655e87775e6d88e48891ffa312cb5326a86be29040a64603993b"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.353500 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d50035b9-6d8a-4760-8088-2a9771cb63d1","Type":"ContainerStarted","Data":"afdc1b01f515be01ff1572f57b769cc2ee00e8458e3283e9b6822130378ef9c3"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.353513 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d50035b9-6d8a-4760-8088-2a9771cb63d1","Type":"ContainerStarted","Data":"5369b74219dae8d64e24935487405583ff1a208f52066c458a5ae60b1156cbeb"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.355266 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"844a3369-74af-4519-919b-f55b202c9e5e","Type":"ContainerStarted","Data":"f94bd8012eb2b360865e3fcec32c031ac08e27157558fef0fb0f2363d64abb8e"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.355305 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"844a3369-74af-4519-919b-f55b202c9e5e","Type":"ContainerStarted","Data":"8545d4237ec997a95346870762acce5a35532a0118a2c311eafc4c948674f1f1"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.355348 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.359514 4733 generic.go:334] "Generic (PLEG): container finished" podID="fffc6200-ffec-4120-9e25-feb4982180c3" containerID="0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2" exitCode=0 Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.359635 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fffc6200-ffec-4120-9e25-feb4982180c3","Type":"ContainerDied","Data":"0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.359666 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fffc6200-ffec-4120-9e25-feb4982180c3","Type":"ContainerDied","Data":"99af1f1058b24405ec7e48d9c9fcfb7105d773e0ed03aa522eb846075f54aa83"} Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.359705 4733 scope.go:117] "RemoveContainer" containerID="0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.360012 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.393750 4733 scope.go:117] "RemoveContainer" containerID="0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2" Dec 04 20:13:17 crc kubenswrapper[4733]: E1204 20:13:17.394362 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2\": container with ID starting with 0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2 not found: ID does not exist" containerID="0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.394404 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2"} err="failed to get container status \"0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2\": rpc error: code = NotFound desc = could not find container \"0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2\": container with ID starting with 0a2575ba0dd3ddd0c3d2d67583cc2277a11d7ce644580c0291a26dc35eb4cfd2 not found: ID does not exist" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.396124 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.396104639 podStartE2EDuration="2.396104639s" podCreationTimestamp="2025-12-04 20:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 20:13:17.367079672 +0000 UTC m=+9259.322440728" watchObservedRunningTime="2025-12-04 20:13:17.396104639 +0000 UTC m=+9259.351465685" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.396449 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.396444518 podStartE2EDuration="2.396444518s" podCreationTimestamp="2025-12-04 20:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 20:13:17.382206953 +0000 UTC m=+9259.337567999" watchObservedRunningTime="2025-12-04 20:13:17.396444518 +0000 UTC m=+9259.351805564" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.416360 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.416342458 podStartE2EDuration="2.416342458s" podCreationTimestamp="2025-12-04 20:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 20:13:17.407630261 +0000 UTC m=+9259.362991327" watchObservedRunningTime="2025-12-04 20:13:17.416342458 +0000 UTC m=+9259.371703504" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.444368 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.460850 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.492786 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 20:13:17 crc kubenswrapper[4733]: E1204 20:13:17.493357 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fffc6200-ffec-4120-9e25-feb4982180c3" containerName="nova-scheduler-scheduler" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.493373 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fffc6200-ffec-4120-9e25-feb4982180c3" containerName="nova-scheduler-scheduler" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.493584 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="fffc6200-ffec-4120-9e25-feb4982180c3" containerName="nova-scheduler-scheduler" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.494451 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.496569 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.513872 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.640284 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-config-data\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.640408 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n4wc\" (UniqueName: \"kubernetes.io/projected/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-kube-api-access-6n4wc\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.640479 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.742538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-config-data\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.742648 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n4wc\" (UniqueName: \"kubernetes.io/projected/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-kube-api-access-6n4wc\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.742703 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.748246 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.765304 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-config-data\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.783559 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n4wc\" (UniqueName: \"kubernetes.io/projected/d1599c17-6cbd-45fe-bfa0-b7862341e8d6-kube-api-access-6n4wc\") pod \"nova-scheduler-0\" (UID: \"d1599c17-6cbd-45fe-bfa0-b7862341e8d6\") " pod="openstack/nova-scheduler-0" Dec 04 20:13:17 crc kubenswrapper[4733]: I1204 20:13:17.814842 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 20:13:18 crc kubenswrapper[4733]: I1204 20:13:18.350505 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fffc6200-ffec-4120-9e25-feb4982180c3" path="/var/lib/kubelet/pods/fffc6200-ffec-4120-9e25-feb4982180c3/volumes" Dec 04 20:13:18 crc kubenswrapper[4733]: I1204 20:13:18.376659 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 20:13:18 crc kubenswrapper[4733]: W1204 20:13:18.393880 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1599c17_6cbd_45fe_bfa0_b7862341e8d6.slice/crio-d74c3b81845d55a397f20102da1311091737afd586c4f84ac8f1bcd2fa36aff9 WatchSource:0}: Error finding container d74c3b81845d55a397f20102da1311091737afd586c4f84ac8f1bcd2fa36aff9: Status 404 returned error can't find the container with id d74c3b81845d55a397f20102da1311091737afd586c4f84ac8f1bcd2fa36aff9 Dec 04 20:13:19 crc kubenswrapper[4733]: I1204 20:13:19.386629 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1599c17-6cbd-45fe-bfa0-b7862341e8d6","Type":"ContainerStarted","Data":"ecee0bd42d7f62416ea0a9e87c6782f11229367fd4eee02d535ca5019f0cbaba"} Dec 04 20:13:19 crc kubenswrapper[4733]: I1204 20:13:19.387273 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1599c17-6cbd-45fe-bfa0-b7862341e8d6","Type":"ContainerStarted","Data":"d74c3b81845d55a397f20102da1311091737afd586c4f84ac8f1bcd2fa36aff9"} Dec 04 20:13:19 crc kubenswrapper[4733]: I1204 20:13:19.412677 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.412651231 podStartE2EDuration="2.412651231s" podCreationTimestamp="2025-12-04 20:13:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 20:13:19.407333057 +0000 UTC m=+9261.362694123" watchObservedRunningTime="2025-12-04 20:13:19.412651231 +0000 UTC m=+9261.368012287" Dec 04 20:13:20 crc kubenswrapper[4733]: I1204 20:13:20.940545 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 20:13:20 crc kubenswrapper[4733]: I1204 20:13:20.940936 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 20:13:22 crc kubenswrapper[4733]: I1204 20:13:22.338006 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:13:22 crc kubenswrapper[4733]: E1204 20:13:22.339449 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:13:22 crc kubenswrapper[4733]: I1204 20:13:22.815605 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 20:13:22 crc kubenswrapper[4733]: I1204 20:13:22.997697 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 20:13:25 crc kubenswrapper[4733]: I1204 20:13:25.823740 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 20:13:25 crc kubenswrapper[4733]: I1204 20:13:25.926709 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 20:13:25 crc kubenswrapper[4733]: I1204 20:13:25.926775 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 20:13:25 crc kubenswrapper[4733]: I1204 20:13:25.940562 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 20:13:25 crc kubenswrapper[4733]: I1204 20:13:25.940871 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.527763 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zg8rt"] Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.534603 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.552371 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zg8rt"] Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.647834 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-utilities\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.647903 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-catalog-content\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.648313 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgcvd\" (UniqueName: \"kubernetes.io/projected/9c02559c-fd79-4704-8bd7-87cc6277512c-kube-api-access-fgcvd\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.750830 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgcvd\" (UniqueName: \"kubernetes.io/projected/9c02559c-fd79-4704-8bd7-87cc6277512c-kube-api-access-fgcvd\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.751062 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-utilities\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.751109 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-catalog-content\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.751595 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-utilities\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.751753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-catalog-content\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.776854 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgcvd\" (UniqueName: \"kubernetes.io/projected/9c02559c-fd79-4704-8bd7-87cc6277512c-kube-api-access-fgcvd\") pod \"certified-operators-zg8rt\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.890615 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:26 crc kubenswrapper[4733]: I1204 20:13:26.974071 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="04fab59a-c08a-4fd3-926a-51a6bd6e6d3f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 20:13:27 crc kubenswrapper[4733]: I1204 20:13:27.022271 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="04fab59a-c08a-4fd3-926a-51a6bd6e6d3f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 20:13:27 crc kubenswrapper[4733]: I1204 20:13:27.108073 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d50035b9-6d8a-4760-8088-2a9771cb63d1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.198:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 20:13:27 crc kubenswrapper[4733]: I1204 20:13:27.109491 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d50035b9-6d8a-4760-8088-2a9771cb63d1" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.198:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 20:13:27 crc kubenswrapper[4733]: I1204 20:13:27.475658 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zg8rt"] Dec 04 20:13:27 crc kubenswrapper[4733]: W1204 20:13:27.483211 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c02559c_fd79_4704_8bd7_87cc6277512c.slice/crio-e9ae484308383755d6c1b62fcd81b827066e6d776b3ede55a8edabdd96b3667e WatchSource:0}: Error finding container e9ae484308383755d6c1b62fcd81b827066e6d776b3ede55a8edabdd96b3667e: Status 404 returned error can't find the container with id e9ae484308383755d6c1b62fcd81b827066e6d776b3ede55a8edabdd96b3667e Dec 04 20:13:27 crc kubenswrapper[4733]: I1204 20:13:27.816327 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 20:13:27 crc kubenswrapper[4733]: I1204 20:13:27.855448 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 20:13:28 crc kubenswrapper[4733]: I1204 20:13:28.502343 4733 generic.go:334] "Generic (PLEG): container finished" podID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerID="be5874291c1de49850b8f1c0380b1d0eae733c1c355a1ed9ebe92faeb8ea811a" exitCode=0 Dec 04 20:13:28 crc kubenswrapper[4733]: I1204 20:13:28.502398 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg8rt" event={"ID":"9c02559c-fd79-4704-8bd7-87cc6277512c","Type":"ContainerDied","Data":"be5874291c1de49850b8f1c0380b1d0eae733c1c355a1ed9ebe92faeb8ea811a"} Dec 04 20:13:28 crc kubenswrapper[4733]: I1204 20:13:28.503460 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg8rt" event={"ID":"9c02559c-fd79-4704-8bd7-87cc6277512c","Type":"ContainerStarted","Data":"e9ae484308383755d6c1b62fcd81b827066e6d776b3ede55a8edabdd96b3667e"} Dec 04 20:13:28 crc kubenswrapper[4733]: I1204 20:13:28.509184 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 20:13:28 crc kubenswrapper[4733]: I1204 20:13:28.564277 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 20:13:30 crc kubenswrapper[4733]: I1204 20:13:30.529074 4733 generic.go:334] "Generic (PLEG): container finished" podID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerID="0e855a766cf68575a7cc24f39241fe8f91980dde7b939ade1ff7e0da3bf888f0" exitCode=0 Dec 04 20:13:30 crc kubenswrapper[4733]: I1204 20:13:30.529954 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg8rt" event={"ID":"9c02559c-fd79-4704-8bd7-87cc6277512c","Type":"ContainerDied","Data":"0e855a766cf68575a7cc24f39241fe8f91980dde7b939ade1ff7e0da3bf888f0"} Dec 04 20:13:32 crc kubenswrapper[4733]: I1204 20:13:32.554332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg8rt" event={"ID":"9c02559c-fd79-4704-8bd7-87cc6277512c","Type":"ContainerStarted","Data":"7e6989849ccebd313015c5992fbf4660f5c31fc922f7af82b22d6052789a59fa"} Dec 04 20:13:32 crc kubenswrapper[4733]: I1204 20:13:32.579085 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zg8rt" podStartSLOduration=4.140129183 podStartE2EDuration="6.579067293s" podCreationTimestamp="2025-12-04 20:13:26 +0000 UTC" firstStartedPulling="2025-12-04 20:13:28.505227674 +0000 UTC m=+9270.460588740" lastFinishedPulling="2025-12-04 20:13:30.944165714 +0000 UTC m=+9272.899526850" observedRunningTime="2025-12-04 20:13:32.57195324 +0000 UTC m=+9274.527314306" watchObservedRunningTime="2025-12-04 20:13:32.579067293 +0000 UTC m=+9274.534428339" Dec 04 20:13:35 crc kubenswrapper[4733]: I1204 20:13:35.930230 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 20:13:35 crc kubenswrapper[4733]: I1204 20:13:35.931763 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 20:13:35 crc kubenswrapper[4733]: I1204 20:13:35.932630 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 20:13:35 crc kubenswrapper[4733]: I1204 20:13:35.936455 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 20:13:35 crc kubenswrapper[4733]: I1204 20:13:35.944370 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 20:13:35 crc kubenswrapper[4733]: I1204 20:13:35.944531 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 20:13:35 crc kubenswrapper[4733]: I1204 20:13:35.949627 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 20:13:35 crc kubenswrapper[4733]: I1204 20:13:35.951378 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 20:13:36 crc kubenswrapper[4733]: I1204 20:13:36.336091 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:13:36 crc kubenswrapper[4733]: E1204 20:13:36.336677 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:13:36 crc kubenswrapper[4733]: I1204 20:13:36.596916 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 20:13:36 crc kubenswrapper[4733]: I1204 20:13:36.600359 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 20:13:36 crc kubenswrapper[4733]: I1204 20:13:36.891406 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:36 crc kubenswrapper[4733]: I1204 20:13:36.891484 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:36 crc kubenswrapper[4733]: I1204 20:13:36.952234 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.576064 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k"] Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.578269 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.581856 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.582220 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.582654 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.582974 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.583179 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fwxxx" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.583326 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.587291 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k"] Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.590694 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.656414 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.707251 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.707312 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.708180 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.708738 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfjhs\" (UniqueName: \"kubernetes.io/projected/5d8da482-a612-4e66-a0f7-7b4f48901cac-kube-api-access-gfjhs\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.708905 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.708965 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.709072 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.709106 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.709191 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.709276 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.709362 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.710630 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zg8rt"] Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.811433 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.811819 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.811888 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.811923 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.812499 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.812545 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.812589 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.812699 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.812743 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.812821 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.812859 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfjhs\" (UniqueName: \"kubernetes.io/projected/5d8da482-a612-4e66-a0f7-7b4f48901cac-kube-api-access-gfjhs\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.813557 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.814036 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.817432 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.818002 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.819855 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.820126 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.820281 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.820468 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.831379 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.832553 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.835487 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfjhs\" (UniqueName: \"kubernetes.io/projected/5d8da482-a612-4e66-a0f7-7b4f48901cac-kube-api-access-gfjhs\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:37 crc kubenswrapper[4733]: I1204 20:13:37.902396 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:13:38 crc kubenswrapper[4733]: W1204 20:13:38.565980 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d8da482_a612_4e66_a0f7_7b4f48901cac.slice/crio-b17cccd33701a99404d84b1b9e68deab759b00e5a1c2b04d44e28e6e8aa9eb6b WatchSource:0}: Error finding container b17cccd33701a99404d84b1b9e68deab759b00e5a1c2b04d44e28e6e8aa9eb6b: Status 404 returned error can't find the container with id b17cccd33701a99404d84b1b9e68deab759b00e5a1c2b04d44e28e6e8aa9eb6b Dec 04 20:13:38 crc kubenswrapper[4733]: I1204 20:13:38.569806 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k"] Dec 04 20:13:38 crc kubenswrapper[4733]: I1204 20:13:38.628357 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" event={"ID":"5d8da482-a612-4e66-a0f7-7b4f48901cac","Type":"ContainerStarted","Data":"b17cccd33701a99404d84b1b9e68deab759b00e5a1c2b04d44e28e6e8aa9eb6b"} Dec 04 20:13:39 crc kubenswrapper[4733]: I1204 20:13:39.640958 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zg8rt" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerName="registry-server" containerID="cri-o://7e6989849ccebd313015c5992fbf4660f5c31fc922f7af82b22d6052789a59fa" gracePeriod=2 Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.656171 4733 generic.go:334] "Generic (PLEG): container finished" podID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerID="7e6989849ccebd313015c5992fbf4660f5c31fc922f7af82b22d6052789a59fa" exitCode=0 Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.656834 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg8rt" event={"ID":"9c02559c-fd79-4704-8bd7-87cc6277512c","Type":"ContainerDied","Data":"7e6989849ccebd313015c5992fbf4660f5c31fc922f7af82b22d6052789a59fa"} Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.656861 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg8rt" event={"ID":"9c02559c-fd79-4704-8bd7-87cc6277512c","Type":"ContainerDied","Data":"e9ae484308383755d6c1b62fcd81b827066e6d776b3ede55a8edabdd96b3667e"} Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.656879 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9ae484308383755d6c1b62fcd81b827066e6d776b3ede55a8edabdd96b3667e" Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.659273 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" event={"ID":"5d8da482-a612-4e66-a0f7-7b4f48901cac","Type":"ContainerStarted","Data":"593d6a93a4cee531559930a580b217a29a85dd186241a02c87b70d68d430668a"} Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.680030 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.681879 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" podStartSLOduration=2.306890107 podStartE2EDuration="3.681696007s" podCreationTimestamp="2025-12-04 20:13:37 +0000 UTC" firstStartedPulling="2025-12-04 20:13:38.568004622 +0000 UTC m=+9280.523365668" lastFinishedPulling="2025-12-04 20:13:39.942810502 +0000 UTC m=+9281.898171568" observedRunningTime="2025-12-04 20:13:40.677880084 +0000 UTC m=+9282.633241140" watchObservedRunningTime="2025-12-04 20:13:40.681696007 +0000 UTC m=+9282.637057053" Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.797321 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgcvd\" (UniqueName: \"kubernetes.io/projected/9c02559c-fd79-4704-8bd7-87cc6277512c-kube-api-access-fgcvd\") pod \"9c02559c-fd79-4704-8bd7-87cc6277512c\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.797435 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-utilities\") pod \"9c02559c-fd79-4704-8bd7-87cc6277512c\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.797543 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-catalog-content\") pod \"9c02559c-fd79-4704-8bd7-87cc6277512c\" (UID: \"9c02559c-fd79-4704-8bd7-87cc6277512c\") " Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.798874 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-utilities" (OuterVolumeSpecName: "utilities") pod "9c02559c-fd79-4704-8bd7-87cc6277512c" (UID: "9c02559c-fd79-4704-8bd7-87cc6277512c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.803024 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c02559c-fd79-4704-8bd7-87cc6277512c-kube-api-access-fgcvd" (OuterVolumeSpecName: "kube-api-access-fgcvd") pod "9c02559c-fd79-4704-8bd7-87cc6277512c" (UID: "9c02559c-fd79-4704-8bd7-87cc6277512c"). InnerVolumeSpecName "kube-api-access-fgcvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.863223 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c02559c-fd79-4704-8bd7-87cc6277512c" (UID: "9c02559c-fd79-4704-8bd7-87cc6277512c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.901696 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgcvd\" (UniqueName: \"kubernetes.io/projected/9c02559c-fd79-4704-8bd7-87cc6277512c-kube-api-access-fgcvd\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.901870 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:40 crc kubenswrapper[4733]: I1204 20:13:40.901963 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02559c-fd79-4704-8bd7-87cc6277512c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:13:41 crc kubenswrapper[4733]: I1204 20:13:41.668468 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg8rt" Dec 04 20:13:41 crc kubenswrapper[4733]: I1204 20:13:41.706196 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zg8rt"] Dec 04 20:13:41 crc kubenswrapper[4733]: I1204 20:13:41.715384 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zg8rt"] Dec 04 20:13:42 crc kubenswrapper[4733]: I1204 20:13:42.349904 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" path="/var/lib/kubelet/pods/9c02559c-fd79-4704-8bd7-87cc6277512c/volumes" Dec 04 20:13:51 crc kubenswrapper[4733]: I1204 20:13:51.335774 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:13:51 crc kubenswrapper[4733]: E1204 20:13:51.336948 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:14:02 crc kubenswrapper[4733]: I1204 20:14:02.338581 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:14:02 crc kubenswrapper[4733]: E1204 20:14:02.339692 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:14:14 crc kubenswrapper[4733]: I1204 20:14:14.336563 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:14:14 crc kubenswrapper[4733]: E1204 20:14:14.337872 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:14:26 crc kubenswrapper[4733]: I1204 20:14:26.336356 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:14:26 crc kubenswrapper[4733]: E1204 20:14:26.337375 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:14:41 crc kubenswrapper[4733]: I1204 20:14:41.336627 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:14:41 crc kubenswrapper[4733]: E1204 20:14:41.338002 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:14:53 crc kubenswrapper[4733]: I1204 20:14:53.335474 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:14:53 crc kubenswrapper[4733]: E1204 20:14:53.336470 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.146493 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5"] Dec 04 20:15:00 crc kubenswrapper[4733]: E1204 20:15:00.147604 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerName="extract-utilities" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.147622 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerName="extract-utilities" Dec 04 20:15:00 crc kubenswrapper[4733]: E1204 20:15:00.147647 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerName="extract-content" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.147657 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerName="extract-content" Dec 04 20:15:00 crc kubenswrapper[4733]: E1204 20:15:00.147681 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerName="registry-server" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.147690 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerName="registry-server" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.148040 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c02559c-fd79-4704-8bd7-87cc6277512c" containerName="registry-server" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.150696 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.170883 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.171196 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.173601 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5"] Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.250126 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wjx6\" (UniqueName: \"kubernetes.io/projected/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-kube-api-access-8wjx6\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.250241 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-config-volume\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.250390 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-secret-volume\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.352492 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wjx6\" (UniqueName: \"kubernetes.io/projected/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-kube-api-access-8wjx6\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.352555 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-config-volume\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.352637 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-secret-volume\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.353688 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-config-volume\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.358267 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-secret-volume\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.380756 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wjx6\" (UniqueName: \"kubernetes.io/projected/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-kube-api-access-8wjx6\") pod \"collect-profiles-29414655-tmgd5\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.506999 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:00 crc kubenswrapper[4733]: I1204 20:15:00.796443 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5"] Dec 04 20:15:01 crc kubenswrapper[4733]: I1204 20:15:01.553640 4733 generic.go:334] "Generic (PLEG): container finished" podID="2bbe0f5d-0180-4ede-a5f0-c2482f0beb06" containerID="675e4fcb6edefc53c0b05333e8e3ba152fb94561174f096a809fe79794786e6b" exitCode=0 Dec 04 20:15:01 crc kubenswrapper[4733]: I1204 20:15:01.553981 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" event={"ID":"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06","Type":"ContainerDied","Data":"675e4fcb6edefc53c0b05333e8e3ba152fb94561174f096a809fe79794786e6b"} Dec 04 20:15:01 crc kubenswrapper[4733]: I1204 20:15:01.554010 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" event={"ID":"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06","Type":"ContainerStarted","Data":"392a78abc01cc3da0b97ffec7c472fcd54b874018817d946fb3caee1499ccdc4"} Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.017730 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.217753 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-config-volume\") pod \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.218175 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-secret-volume\") pod \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.218248 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wjx6\" (UniqueName: \"kubernetes.io/projected/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-kube-api-access-8wjx6\") pod \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\" (UID: \"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06\") " Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.218733 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-config-volume" (OuterVolumeSpecName: "config-volume") pod "2bbe0f5d-0180-4ede-a5f0-c2482f0beb06" (UID: "2bbe0f5d-0180-4ede-a5f0-c2482f0beb06"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.226836 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2bbe0f5d-0180-4ede-a5f0-c2482f0beb06" (UID: "2bbe0f5d-0180-4ede-a5f0-c2482f0beb06"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.229994 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-kube-api-access-8wjx6" (OuterVolumeSpecName: "kube-api-access-8wjx6") pod "2bbe0f5d-0180-4ede-a5f0-c2482f0beb06" (UID: "2bbe0f5d-0180-4ede-a5f0-c2482f0beb06"). InnerVolumeSpecName "kube-api-access-8wjx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.320405 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.320453 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.320473 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wjx6\" (UniqueName: \"kubernetes.io/projected/2bbe0f5d-0180-4ede-a5f0-c2482f0beb06-kube-api-access-8wjx6\") on node \"crc\" DevicePath \"\"" Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.576892 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" event={"ID":"2bbe0f5d-0180-4ede-a5f0-c2482f0beb06","Type":"ContainerDied","Data":"392a78abc01cc3da0b97ffec7c472fcd54b874018817d946fb3caee1499ccdc4"} Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.576928 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="392a78abc01cc3da0b97ffec7c472fcd54b874018817d946fb3caee1499ccdc4" Dec 04 20:15:03 crc kubenswrapper[4733]: I1204 20:15:03.576972 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414655-tmgd5" Dec 04 20:15:04 crc kubenswrapper[4733]: I1204 20:15:04.102583 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs"] Dec 04 20:15:04 crc kubenswrapper[4733]: I1204 20:15:04.112786 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414610-wx5vs"] Dec 04 20:15:04 crc kubenswrapper[4733]: I1204 20:15:04.335852 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:15:04 crc kubenswrapper[4733]: E1204 20:15:04.336397 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:15:04 crc kubenswrapper[4733]: I1204 20:15:04.351932 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91fcf596-4fd0-4332-99e7-376ae3dfac29" path="/var/lib/kubelet/pods/91fcf596-4fd0-4332-99e7-376ae3dfac29/volumes" Dec 04 20:15:15 crc kubenswrapper[4733]: I1204 20:15:15.336246 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:15:15 crc kubenswrapper[4733]: E1204 20:15:15.337329 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:15:28 crc kubenswrapper[4733]: I1204 20:15:28.345191 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:15:28 crc kubenswrapper[4733]: E1204 20:15:28.350420 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:15:29 crc kubenswrapper[4733]: I1204 20:15:29.825695 4733 scope.go:117] "RemoveContainer" containerID="5ad51377fd89bff828aeb2e6411eaba6e29b3e3bd1b51c2e8be3320cbeda7f88" Dec 04 20:15:40 crc kubenswrapper[4733]: I1204 20:15:40.336007 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:15:40 crc kubenswrapper[4733]: E1204 20:15:40.337480 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:15:54 crc kubenswrapper[4733]: I1204 20:15:54.336211 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:15:54 crc kubenswrapper[4733]: E1204 20:15:54.338065 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:16:05 crc kubenswrapper[4733]: I1204 20:16:05.336464 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:16:05 crc kubenswrapper[4733]: E1204 20:16:05.337549 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:16:16 crc kubenswrapper[4733]: I1204 20:16:16.335937 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:16:16 crc kubenswrapper[4733]: E1204 20:16:16.336748 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:16:29 crc kubenswrapper[4733]: I1204 20:16:29.335447 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:16:29 crc kubenswrapper[4733]: E1204 20:16:29.336275 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:16:42 crc kubenswrapper[4733]: I1204 20:16:42.337424 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:16:42 crc kubenswrapper[4733]: E1204 20:16:42.338352 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:16:53 crc kubenswrapper[4733]: I1204 20:16:53.336914 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:16:53 crc kubenswrapper[4733]: E1204 20:16:53.337684 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:17:08 crc kubenswrapper[4733]: I1204 20:17:08.350559 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:17:08 crc kubenswrapper[4733]: E1204 20:17:08.351291 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:17:20 crc kubenswrapper[4733]: I1204 20:17:20.335868 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:17:20 crc kubenswrapper[4733]: E1204 20:17:20.336655 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:17:33 crc kubenswrapper[4733]: I1204 20:17:33.336038 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:17:33 crc kubenswrapper[4733]: E1204 20:17:33.337152 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:17:41 crc kubenswrapper[4733]: I1204 20:17:41.802177 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vrbwk"] Dec 04 20:17:41 crc kubenswrapper[4733]: E1204 20:17:41.803325 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbe0f5d-0180-4ede-a5f0-c2482f0beb06" containerName="collect-profiles" Dec 04 20:17:41 crc kubenswrapper[4733]: I1204 20:17:41.803342 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbe0f5d-0180-4ede-a5f0-c2482f0beb06" containerName="collect-profiles" Dec 04 20:17:41 crc kubenswrapper[4733]: I1204 20:17:41.803678 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbe0f5d-0180-4ede-a5f0-c2482f0beb06" containerName="collect-profiles" Dec 04 20:17:41 crc kubenswrapper[4733]: I1204 20:17:41.806197 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:41 crc kubenswrapper[4733]: I1204 20:17:41.820702 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrbwk"] Dec 04 20:17:41 crc kubenswrapper[4733]: I1204 20:17:41.943180 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgbpq\" (UniqueName: \"kubernetes.io/projected/c71f034e-d405-4d52-8fae-91a11955a627-kube-api-access-rgbpq\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:41 crc kubenswrapper[4733]: I1204 20:17:41.943274 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-utilities\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:41 crc kubenswrapper[4733]: I1204 20:17:41.943319 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-catalog-content\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:42 crc kubenswrapper[4733]: I1204 20:17:42.045808 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-utilities\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:42 crc kubenswrapper[4733]: I1204 20:17:42.046000 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-catalog-content\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:42 crc kubenswrapper[4733]: I1204 20:17:42.046364 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-utilities\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:42 crc kubenswrapper[4733]: I1204 20:17:42.046530 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-catalog-content\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:42 crc kubenswrapper[4733]: I1204 20:17:42.046659 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgbpq\" (UniqueName: \"kubernetes.io/projected/c71f034e-d405-4d52-8fae-91a11955a627-kube-api-access-rgbpq\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:42 crc kubenswrapper[4733]: I1204 20:17:42.068217 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgbpq\" (UniqueName: \"kubernetes.io/projected/c71f034e-d405-4d52-8fae-91a11955a627-kube-api-access-rgbpq\") pod \"community-operators-vrbwk\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:42 crc kubenswrapper[4733]: I1204 20:17:42.159390 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:42 crc kubenswrapper[4733]: I1204 20:17:42.768643 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrbwk"] Dec 04 20:17:42 crc kubenswrapper[4733]: W1204 20:17:42.775115 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc71f034e_d405_4d52_8fae_91a11955a627.slice/crio-f5b2686b06b5d222e70c099cb5c2811f57a04aed730ebf000bd223400837bb43 WatchSource:0}: Error finding container f5b2686b06b5d222e70c099cb5c2811f57a04aed730ebf000bd223400837bb43: Status 404 returned error can't find the container with id f5b2686b06b5d222e70c099cb5c2811f57a04aed730ebf000bd223400837bb43 Dec 04 20:17:43 crc kubenswrapper[4733]: I1204 20:17:43.535987 4733 generic.go:334] "Generic (PLEG): container finished" podID="c71f034e-d405-4d52-8fae-91a11955a627" containerID="aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9" exitCode=0 Dec 04 20:17:43 crc kubenswrapper[4733]: I1204 20:17:43.536116 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbwk" event={"ID":"c71f034e-d405-4d52-8fae-91a11955a627","Type":"ContainerDied","Data":"aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9"} Dec 04 20:17:43 crc kubenswrapper[4733]: I1204 20:17:43.536373 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbwk" event={"ID":"c71f034e-d405-4d52-8fae-91a11955a627","Type":"ContainerStarted","Data":"f5b2686b06b5d222e70c099cb5c2811f57a04aed730ebf000bd223400837bb43"} Dec 04 20:17:44 crc kubenswrapper[4733]: I1204 20:17:44.335360 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:17:44 crc kubenswrapper[4733]: E1204 20:17:44.336125 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:17:44 crc kubenswrapper[4733]: I1204 20:17:44.551278 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbwk" event={"ID":"c71f034e-d405-4d52-8fae-91a11955a627","Type":"ContainerStarted","Data":"f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714"} Dec 04 20:17:45 crc kubenswrapper[4733]: I1204 20:17:45.569442 4733 generic.go:334] "Generic (PLEG): container finished" podID="c71f034e-d405-4d52-8fae-91a11955a627" containerID="f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714" exitCode=0 Dec 04 20:17:45 crc kubenswrapper[4733]: I1204 20:17:45.569496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbwk" event={"ID":"c71f034e-d405-4d52-8fae-91a11955a627","Type":"ContainerDied","Data":"f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714"} Dec 04 20:17:47 crc kubenswrapper[4733]: I1204 20:17:47.591524 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbwk" event={"ID":"c71f034e-d405-4d52-8fae-91a11955a627","Type":"ContainerStarted","Data":"2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f"} Dec 04 20:17:47 crc kubenswrapper[4733]: I1204 20:17:47.618450 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vrbwk" podStartSLOduration=4.17788792 podStartE2EDuration="6.618432413s" podCreationTimestamp="2025-12-04 20:17:41 +0000 UTC" firstStartedPulling="2025-12-04 20:17:43.538673925 +0000 UTC m=+9525.494035011" lastFinishedPulling="2025-12-04 20:17:45.979218448 +0000 UTC m=+9527.934579504" observedRunningTime="2025-12-04 20:17:47.617063666 +0000 UTC m=+9529.572424722" watchObservedRunningTime="2025-12-04 20:17:47.618432413 +0000 UTC m=+9529.573793469" Dec 04 20:17:49 crc kubenswrapper[4733]: I1204 20:17:49.612171 4733 generic.go:334] "Generic (PLEG): container finished" podID="5d8da482-a612-4e66-a0f7-7b4f48901cac" containerID="593d6a93a4cee531559930a580b217a29a85dd186241a02c87b70d68d430668a" exitCode=0 Dec 04 20:17:49 crc kubenswrapper[4733]: I1204 20:17:49.612257 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" event={"ID":"5d8da482-a612-4e66-a0f7-7b4f48901cac","Type":"ContainerDied","Data":"593d6a93a4cee531559930a580b217a29a85dd186241a02c87b70d68d430668a"} Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.166992 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276403 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-0\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276505 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ceph\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276576 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-0\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276675 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-1\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276745 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfjhs\" (UniqueName: \"kubernetes.io/projected/5d8da482-a612-4e66-a0f7-7b4f48901cac-kube-api-access-gfjhs\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276782 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-combined-ca-bundle\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276836 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-1\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276859 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ssh-key\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276884 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-1\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276951 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-inventory\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.276975 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-0\") pod \"5d8da482-a612-4e66-a0f7-7b4f48901cac\" (UID: \"5d8da482-a612-4e66-a0f7-7b4f48901cac\") " Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.283005 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.283342 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d8da482-a612-4e66-a0f7-7b4f48901cac-kube-api-access-gfjhs" (OuterVolumeSpecName: "kube-api-access-gfjhs") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "kube-api-access-gfjhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.285869 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ceph" (OuterVolumeSpecName: "ceph") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.306127 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.309736 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.312160 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.314447 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-inventory" (OuterVolumeSpecName: "inventory") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.317758 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.318065 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.319482 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.320417 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d8da482-a612-4e66-a0f7-7b4f48901cac" (UID: "5d8da482-a612-4e66-a0f7-7b4f48901cac"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379341 4733 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379377 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfjhs\" (UniqueName: \"kubernetes.io/projected/5d8da482-a612-4e66-a0f7-7b4f48901cac-kube-api-access-gfjhs\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379386 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379395 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379405 4733 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379416 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379424 4733 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379433 4733 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379441 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379448 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.379456 4733 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5d8da482-a612-4e66-a0f7-7b4f48901cac-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.645537 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" event={"ID":"5d8da482-a612-4e66-a0f7-7b4f48901cac","Type":"ContainerDied","Data":"b17cccd33701a99404d84b1b9e68deab759b00e5a1c2b04d44e28e6e8aa9eb6b"} Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.645585 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b17cccd33701a99404d84b1b9e68deab759b00e5a1c2b04d44e28e6e8aa9eb6b" Dec 04 20:17:51 crc kubenswrapper[4733]: I1204 20:17:51.645865 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k" Dec 04 20:17:52 crc kubenswrapper[4733]: I1204 20:17:52.169022 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:52 crc kubenswrapper[4733]: I1204 20:17:52.169093 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:52 crc kubenswrapper[4733]: I1204 20:17:52.258501 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:52 crc kubenswrapper[4733]: I1204 20:17:52.751346 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:52 crc kubenswrapper[4733]: I1204 20:17:52.829161 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrbwk"] Dec 04 20:17:54 crc kubenswrapper[4733]: I1204 20:17:54.694442 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vrbwk" podUID="c71f034e-d405-4d52-8fae-91a11955a627" containerName="registry-server" containerID="cri-o://2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f" gracePeriod=2 Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.185503 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.280402 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-catalog-content\") pod \"c71f034e-d405-4d52-8fae-91a11955a627\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.281083 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-utilities\") pod \"c71f034e-d405-4d52-8fae-91a11955a627\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.281218 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgbpq\" (UniqueName: \"kubernetes.io/projected/c71f034e-d405-4d52-8fae-91a11955a627-kube-api-access-rgbpq\") pod \"c71f034e-d405-4d52-8fae-91a11955a627\" (UID: \"c71f034e-d405-4d52-8fae-91a11955a627\") " Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.285173 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-utilities" (OuterVolumeSpecName: "utilities") pod "c71f034e-d405-4d52-8fae-91a11955a627" (UID: "c71f034e-d405-4d52-8fae-91a11955a627"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.290088 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c71f034e-d405-4d52-8fae-91a11955a627-kube-api-access-rgbpq" (OuterVolumeSpecName: "kube-api-access-rgbpq") pod "c71f034e-d405-4d52-8fae-91a11955a627" (UID: "c71f034e-d405-4d52-8fae-91a11955a627"). InnerVolumeSpecName "kube-api-access-rgbpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.350399 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c71f034e-d405-4d52-8fae-91a11955a627" (UID: "c71f034e-d405-4d52-8fae-91a11955a627"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.384140 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.384178 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c71f034e-d405-4d52-8fae-91a11955a627-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.384189 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgbpq\" (UniqueName: \"kubernetes.io/projected/c71f034e-d405-4d52-8fae-91a11955a627-kube-api-access-rgbpq\") on node \"crc\" DevicePath \"\"" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.706612 4733 generic.go:334] "Generic (PLEG): container finished" podID="c71f034e-d405-4d52-8fae-91a11955a627" containerID="2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f" exitCode=0 Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.706655 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbwk" event={"ID":"c71f034e-d405-4d52-8fae-91a11955a627","Type":"ContainerDied","Data":"2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f"} Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.706688 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrbwk" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.706711 4733 scope.go:117] "RemoveContainer" containerID="2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.706698 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrbwk" event={"ID":"c71f034e-d405-4d52-8fae-91a11955a627","Type":"ContainerDied","Data":"f5b2686b06b5d222e70c099cb5c2811f57a04aed730ebf000bd223400837bb43"} Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.731826 4733 scope.go:117] "RemoveContainer" containerID="f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.750494 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrbwk"] Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.758082 4733 scope.go:117] "RemoveContainer" containerID="aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.762152 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vrbwk"] Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.808043 4733 scope.go:117] "RemoveContainer" containerID="2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f" Dec 04 20:17:55 crc kubenswrapper[4733]: E1204 20:17:55.808554 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f\": container with ID starting with 2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f not found: ID does not exist" containerID="2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.808707 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f"} err="failed to get container status \"2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f\": rpc error: code = NotFound desc = could not find container \"2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f\": container with ID starting with 2ff5465afd7caaa7717ff31332bf48e490a37c34fed08536e23df31a3d8efc6f not found: ID does not exist" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.808812 4733 scope.go:117] "RemoveContainer" containerID="f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714" Dec 04 20:17:55 crc kubenswrapper[4733]: E1204 20:17:55.809153 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714\": container with ID starting with f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714 not found: ID does not exist" containerID="f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.809198 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714"} err="failed to get container status \"f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714\": rpc error: code = NotFound desc = could not find container \"f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714\": container with ID starting with f8d53e40528ee2cea872e6226b099806f83901aafad6f5b70261c8adba217714 not found: ID does not exist" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.809247 4733 scope.go:117] "RemoveContainer" containerID="aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9" Dec 04 20:17:55 crc kubenswrapper[4733]: E1204 20:17:55.809662 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9\": container with ID starting with aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9 not found: ID does not exist" containerID="aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9" Dec 04 20:17:55 crc kubenswrapper[4733]: I1204 20:17:55.809834 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9"} err="failed to get container status \"aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9\": rpc error: code = NotFound desc = could not find container \"aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9\": container with ID starting with aa8e048876393ade2d0167434db8f4da35c1dd7456e0807dacfe21571c93e7f9 not found: ID does not exist" Dec 04 20:17:56 crc kubenswrapper[4733]: I1204 20:17:56.349554 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c71f034e-d405-4d52-8fae-91a11955a627" path="/var/lib/kubelet/pods/c71f034e-d405-4d52-8fae-91a11955a627/volumes" Dec 04 20:17:58 crc kubenswrapper[4733]: I1204 20:17:58.347019 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:17:58 crc kubenswrapper[4733]: I1204 20:17:58.748455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"2ae00648d3f84b6e62e8b996f069c0f30a14ec5ac0a0886f49c65f077b2f9bda"} Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.762405 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pgmnc"] Dec 04 20:18:26 crc kubenswrapper[4733]: E1204 20:18:26.765166 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d8da482-a612-4e66-a0f7-7b4f48901cac" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.765307 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d8da482-a612-4e66-a0f7-7b4f48901cac" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 04 20:18:26 crc kubenswrapper[4733]: E1204 20:18:26.765396 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71f034e-d405-4d52-8fae-91a11955a627" containerName="extract-utilities" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.765481 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71f034e-d405-4d52-8fae-91a11955a627" containerName="extract-utilities" Dec 04 20:18:26 crc kubenswrapper[4733]: E1204 20:18:26.765558 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71f034e-d405-4d52-8fae-91a11955a627" containerName="extract-content" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.765628 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71f034e-d405-4d52-8fae-91a11955a627" containerName="extract-content" Dec 04 20:18:26 crc kubenswrapper[4733]: E1204 20:18:26.765732 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71f034e-d405-4d52-8fae-91a11955a627" containerName="registry-server" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.765891 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71f034e-d405-4d52-8fae-91a11955a627" containerName="registry-server" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.766241 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c71f034e-d405-4d52-8fae-91a11955a627" containerName="registry-server" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.766366 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d8da482-a612-4e66-a0f7-7b4f48901cac" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.768569 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.795281 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pgmnc"] Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.901225 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-utilities\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.901291 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-catalog-content\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:26 crc kubenswrapper[4733]: I1204 20:18:26.901382 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8lbj\" (UniqueName: \"kubernetes.io/projected/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-kube-api-access-k8lbj\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:27 crc kubenswrapper[4733]: I1204 20:18:27.003996 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8lbj\" (UniqueName: \"kubernetes.io/projected/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-kube-api-access-k8lbj\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:27 crc kubenswrapper[4733]: I1204 20:18:27.004440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-utilities\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:27 crc kubenswrapper[4733]: I1204 20:18:27.004559 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-catalog-content\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:27 crc kubenswrapper[4733]: I1204 20:18:27.005124 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-utilities\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:27 crc kubenswrapper[4733]: I1204 20:18:27.005330 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-catalog-content\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:27 crc kubenswrapper[4733]: I1204 20:18:27.029850 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8lbj\" (UniqueName: \"kubernetes.io/projected/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-kube-api-access-k8lbj\") pod \"redhat-operators-pgmnc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:27 crc kubenswrapper[4733]: I1204 20:18:27.113599 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:27 crc kubenswrapper[4733]: I1204 20:18:27.593580 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pgmnc"] Dec 04 20:18:28 crc kubenswrapper[4733]: I1204 20:18:28.134003 4733 generic.go:334] "Generic (PLEG): container finished" podID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerID="56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3" exitCode=0 Dec 04 20:18:28 crc kubenswrapper[4733]: I1204 20:18:28.134056 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgmnc" event={"ID":"0a9c1f24-c4bb-4bac-9a77-49325247b5fc","Type":"ContainerDied","Data":"56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3"} Dec 04 20:18:28 crc kubenswrapper[4733]: I1204 20:18:28.134337 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgmnc" event={"ID":"0a9c1f24-c4bb-4bac-9a77-49325247b5fc","Type":"ContainerStarted","Data":"228d15be87ab5bf3115fe1e742369e0e430696f3fcb0a4de73fe9b651c587f0a"} Dec 04 20:18:30 crc kubenswrapper[4733]: I1204 20:18:30.164000 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgmnc" event={"ID":"0a9c1f24-c4bb-4bac-9a77-49325247b5fc","Type":"ContainerStarted","Data":"85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009"} Dec 04 20:18:32 crc kubenswrapper[4733]: I1204 20:18:32.189997 4733 generic.go:334] "Generic (PLEG): container finished" podID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerID="85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009" exitCode=0 Dec 04 20:18:32 crc kubenswrapper[4733]: I1204 20:18:32.190047 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgmnc" event={"ID":"0a9c1f24-c4bb-4bac-9a77-49325247b5fc","Type":"ContainerDied","Data":"85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009"} Dec 04 20:18:32 crc kubenswrapper[4733]: I1204 20:18:32.192690 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 20:18:33 crc kubenswrapper[4733]: I1204 20:18:33.199877 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgmnc" event={"ID":"0a9c1f24-c4bb-4bac-9a77-49325247b5fc","Type":"ContainerStarted","Data":"2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1"} Dec 04 20:18:33 crc kubenswrapper[4733]: I1204 20:18:33.238276 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pgmnc" podStartSLOduration=2.781931075 podStartE2EDuration="7.238255489s" podCreationTimestamp="2025-12-04 20:18:26 +0000 UTC" firstStartedPulling="2025-12-04 20:18:28.135882685 +0000 UTC m=+9570.091243731" lastFinishedPulling="2025-12-04 20:18:32.592207099 +0000 UTC m=+9574.547568145" observedRunningTime="2025-12-04 20:18:33.223487378 +0000 UTC m=+9575.178848424" watchObservedRunningTime="2025-12-04 20:18:33.238255489 +0000 UTC m=+9575.193616535" Dec 04 20:18:37 crc kubenswrapper[4733]: I1204 20:18:37.114474 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:37 crc kubenswrapper[4733]: I1204 20:18:37.115201 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:38 crc kubenswrapper[4733]: I1204 20:18:38.168286 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pgmnc" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="registry-server" probeResult="failure" output=< Dec 04 20:18:38 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 20:18:38 crc kubenswrapper[4733]: > Dec 04 20:18:47 crc kubenswrapper[4733]: I1204 20:18:47.306247 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:47 crc kubenswrapper[4733]: I1204 20:18:47.395155 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:47 crc kubenswrapper[4733]: I1204 20:18:47.550543 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pgmnc"] Dec 04 20:18:48 crc kubenswrapper[4733]: I1204 20:18:48.427566 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pgmnc" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="registry-server" containerID="cri-o://2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1" gracePeriod=2 Dec 04 20:18:48 crc kubenswrapper[4733]: I1204 20:18:48.937043 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.090438 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-utilities\") pod \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.090540 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8lbj\" (UniqueName: \"kubernetes.io/projected/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-kube-api-access-k8lbj\") pod \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.090614 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-catalog-content\") pod \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\" (UID: \"0a9c1f24-c4bb-4bac-9a77-49325247b5fc\") " Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.091296 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-utilities" (OuterVolumeSpecName: "utilities") pod "0a9c1f24-c4bb-4bac-9a77-49325247b5fc" (UID: "0a9c1f24-c4bb-4bac-9a77-49325247b5fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.097302 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-kube-api-access-k8lbj" (OuterVolumeSpecName: "kube-api-access-k8lbj") pod "0a9c1f24-c4bb-4bac-9a77-49325247b5fc" (UID: "0a9c1f24-c4bb-4bac-9a77-49325247b5fc"). InnerVolumeSpecName "kube-api-access-k8lbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.193860 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8lbj\" (UniqueName: \"kubernetes.io/projected/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-kube-api-access-k8lbj\") on node \"crc\" DevicePath \"\"" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.193925 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.203222 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a9c1f24-c4bb-4bac-9a77-49325247b5fc" (UID: "0a9c1f24-c4bb-4bac-9a77-49325247b5fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.296118 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9c1f24-c4bb-4bac-9a77-49325247b5fc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.448338 4733 generic.go:334] "Generic (PLEG): container finished" podID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerID="2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1" exitCode=0 Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.448454 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgmnc" event={"ID":"0a9c1f24-c4bb-4bac-9a77-49325247b5fc","Type":"ContainerDied","Data":"2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1"} Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.449298 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgmnc" event={"ID":"0a9c1f24-c4bb-4bac-9a77-49325247b5fc","Type":"ContainerDied","Data":"228d15be87ab5bf3115fe1e742369e0e430696f3fcb0a4de73fe9b651c587f0a"} Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.449370 4733 scope.go:117] "RemoveContainer" containerID="2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.448479 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgmnc" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.477963 4733 scope.go:117] "RemoveContainer" containerID="85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.496390 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pgmnc"] Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.509670 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pgmnc"] Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.509718 4733 scope.go:117] "RemoveContainer" containerID="56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.549054 4733 scope.go:117] "RemoveContainer" containerID="2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1" Dec 04 20:18:49 crc kubenswrapper[4733]: E1204 20:18:49.550270 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1\": container with ID starting with 2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1 not found: ID does not exist" containerID="2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.550300 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1"} err="failed to get container status \"2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1\": rpc error: code = NotFound desc = could not find container \"2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1\": container with ID starting with 2ad04921be4f49aaca13920828d8fe0eaa5da71ab60008a7cebd3c6b00d5c7b1 not found: ID does not exist" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.550320 4733 scope.go:117] "RemoveContainer" containerID="85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009" Dec 04 20:18:49 crc kubenswrapper[4733]: E1204 20:18:49.550665 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009\": container with ID starting with 85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009 not found: ID does not exist" containerID="85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.550725 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009"} err="failed to get container status \"85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009\": rpc error: code = NotFound desc = could not find container \"85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009\": container with ID starting with 85cd77e4960b2f99c999bff2b7b881bf610b7a5d0e392083c6bc9d820d3d3009 not found: ID does not exist" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.550764 4733 scope.go:117] "RemoveContainer" containerID="56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3" Dec 04 20:18:49 crc kubenswrapper[4733]: E1204 20:18:49.551052 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3\": container with ID starting with 56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3 not found: ID does not exist" containerID="56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3" Dec 04 20:18:49 crc kubenswrapper[4733]: I1204 20:18:49.551075 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3"} err="failed to get container status \"56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3\": rpc error: code = NotFound desc = could not find container \"56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3\": container with ID starting with 56d597eeb5fb0eda0a18fd1f55d735411850d97fa03b8268f09a53d8eb17ede3 not found: ID does not exist" Dec 04 20:18:50 crc kubenswrapper[4733]: I1204 20:18:50.350362 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" path="/var/lib/kubelet/pods/0a9c1f24-c4bb-4bac-9a77-49325247b5fc/volumes" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.243607 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-52qf2"] Dec 04 20:19:28 crc kubenswrapper[4733]: E1204 20:19:28.244579 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="extract-utilities" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.244594 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="extract-utilities" Dec 04 20:19:28 crc kubenswrapper[4733]: E1204 20:19:28.244617 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="extract-content" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.244623 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="extract-content" Dec 04 20:19:28 crc kubenswrapper[4733]: E1204 20:19:28.244648 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="registry-server" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.244655 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="registry-server" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.244891 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9c1f24-c4bb-4bac-9a77-49325247b5fc" containerName="registry-server" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.246414 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.260186 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-52qf2"] Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.388531 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4gzj\" (UniqueName: \"kubernetes.io/projected/34d91348-64d0-44c3-82c9-c5bbacf4271b-kube-api-access-f4gzj\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.388927 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-utilities\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.389218 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-catalog-content\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.491122 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4gzj\" (UniqueName: \"kubernetes.io/projected/34d91348-64d0-44c3-82c9-c5bbacf4271b-kube-api-access-f4gzj\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.491212 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-utilities\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.491323 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-catalog-content\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.491748 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-utilities\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.491778 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-catalog-content\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.514637 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4gzj\" (UniqueName: \"kubernetes.io/projected/34d91348-64d0-44c3-82c9-c5bbacf4271b-kube-api-access-f4gzj\") pod \"redhat-marketplace-52qf2\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:28 crc kubenswrapper[4733]: I1204 20:19:28.567606 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:29 crc kubenswrapper[4733]: I1204 20:19:29.046945 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-52qf2"] Dec 04 20:19:29 crc kubenswrapper[4733]: I1204 20:19:29.920149 4733 generic.go:334] "Generic (PLEG): container finished" podID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerID="f70a266c02b8e3fa0e77f5393bf0ef36f89ba699505cd0104b3c22b791eff081" exitCode=0 Dec 04 20:19:29 crc kubenswrapper[4733]: I1204 20:19:29.920189 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52qf2" event={"ID":"34d91348-64d0-44c3-82c9-c5bbacf4271b","Type":"ContainerDied","Data":"f70a266c02b8e3fa0e77f5393bf0ef36f89ba699505cd0104b3c22b791eff081"} Dec 04 20:19:29 crc kubenswrapper[4733]: I1204 20:19:29.920215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52qf2" event={"ID":"34d91348-64d0-44c3-82c9-c5bbacf4271b","Type":"ContainerStarted","Data":"5e5817c3c282187d49227f258eebdd81baaca73fcbb41a6f4caa23cee01c9fed"} Dec 04 20:19:29 crc kubenswrapper[4733]: I1204 20:19:29.975686 4733 scope.go:117] "RemoveContainer" containerID="0e855a766cf68575a7cc24f39241fe8f91980dde7b939ade1ff7e0da3bf888f0" Dec 04 20:19:30 crc kubenswrapper[4733]: I1204 20:19:30.008429 4733 scope.go:117] "RemoveContainer" containerID="be5874291c1de49850b8f1c0380b1d0eae733c1c355a1ed9ebe92faeb8ea811a" Dec 04 20:19:31 crc kubenswrapper[4733]: I1204 20:19:31.948312 4733 generic.go:334] "Generic (PLEG): container finished" podID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerID="524b989bae1bd4467a8b331e0b0b0a5fb8f4bce3cb713f5c3cfae086faa35475" exitCode=0 Dec 04 20:19:31 crc kubenswrapper[4733]: I1204 20:19:31.949080 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52qf2" event={"ID":"34d91348-64d0-44c3-82c9-c5bbacf4271b","Type":"ContainerDied","Data":"524b989bae1bd4467a8b331e0b0b0a5fb8f4bce3cb713f5c3cfae086faa35475"} Dec 04 20:19:32 crc kubenswrapper[4733]: I1204 20:19:32.962229 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52qf2" event={"ID":"34d91348-64d0-44c3-82c9-c5bbacf4271b","Type":"ContainerStarted","Data":"b825261551734f7768c0ff3465fa00437c240137810eee029b584beb03ebbfb2"} Dec 04 20:19:32 crc kubenswrapper[4733]: I1204 20:19:32.987552 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-52qf2" podStartSLOduration=2.544426292 podStartE2EDuration="4.987535865s" podCreationTimestamp="2025-12-04 20:19:28 +0000 UTC" firstStartedPulling="2025-12-04 20:19:29.922194039 +0000 UTC m=+9631.877555095" lastFinishedPulling="2025-12-04 20:19:32.365303632 +0000 UTC m=+9634.320664668" observedRunningTime="2025-12-04 20:19:32.976947238 +0000 UTC m=+9634.932308314" watchObservedRunningTime="2025-12-04 20:19:32.987535865 +0000 UTC m=+9634.942896911" Dec 04 20:19:38 crc kubenswrapper[4733]: I1204 20:19:38.568315 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:38 crc kubenswrapper[4733]: I1204 20:19:38.568922 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:38 crc kubenswrapper[4733]: I1204 20:19:38.615819 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:39 crc kubenswrapper[4733]: I1204 20:19:39.074960 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:39 crc kubenswrapper[4733]: I1204 20:19:39.155016 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-52qf2"] Dec 04 20:19:41 crc kubenswrapper[4733]: I1204 20:19:41.056151 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-52qf2" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerName="registry-server" containerID="cri-o://b825261551734f7768c0ff3465fa00437c240137810eee029b584beb03ebbfb2" gracePeriod=2 Dec 04 20:19:41 crc kubenswrapper[4733]: E1204 20:19:41.213787 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34d91348_64d0_44c3_82c9_c5bbacf4271b.slice/crio-b825261551734f7768c0ff3465fa00437c240137810eee029b584beb03ebbfb2.scope\": RecentStats: unable to find data in memory cache]" Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.074865 4733 generic.go:334] "Generic (PLEG): container finished" podID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerID="b825261551734f7768c0ff3465fa00437c240137810eee029b584beb03ebbfb2" exitCode=0 Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.075063 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52qf2" event={"ID":"34d91348-64d0-44c3-82c9-c5bbacf4271b","Type":"ContainerDied","Data":"b825261551734f7768c0ff3465fa00437c240137810eee029b584beb03ebbfb2"} Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.075590 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52qf2" event={"ID":"34d91348-64d0-44c3-82c9-c5bbacf4271b","Type":"ContainerDied","Data":"5e5817c3c282187d49227f258eebdd81baaca73fcbb41a6f4caa23cee01c9fed"} Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.075613 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e5817c3c282187d49227f258eebdd81baaca73fcbb41a6f4caa23cee01c9fed" Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.175717 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.292713 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4gzj\" (UniqueName: \"kubernetes.io/projected/34d91348-64d0-44c3-82c9-c5bbacf4271b-kube-api-access-f4gzj\") pod \"34d91348-64d0-44c3-82c9-c5bbacf4271b\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.292822 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-utilities\") pod \"34d91348-64d0-44c3-82c9-c5bbacf4271b\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.293067 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-catalog-content\") pod \"34d91348-64d0-44c3-82c9-c5bbacf4271b\" (UID: \"34d91348-64d0-44c3-82c9-c5bbacf4271b\") " Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.294929 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-utilities" (OuterVolumeSpecName: "utilities") pod "34d91348-64d0-44c3-82c9-c5bbacf4271b" (UID: "34d91348-64d0-44c3-82c9-c5bbacf4271b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.302393 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34d91348-64d0-44c3-82c9-c5bbacf4271b-kube-api-access-f4gzj" (OuterVolumeSpecName: "kube-api-access-f4gzj") pod "34d91348-64d0-44c3-82c9-c5bbacf4271b" (UID: "34d91348-64d0-44c3-82c9-c5bbacf4271b"). InnerVolumeSpecName "kube-api-access-f4gzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.315237 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34d91348-64d0-44c3-82c9-c5bbacf4271b" (UID: "34d91348-64d0-44c3-82c9-c5bbacf4271b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.395767 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.396009 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4gzj\" (UniqueName: \"kubernetes.io/projected/34d91348-64d0-44c3-82c9-c5bbacf4271b-kube-api-access-f4gzj\") on node \"crc\" DevicePath \"\"" Dec 04 20:19:42 crc kubenswrapper[4733]: I1204 20:19:42.396073 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34d91348-64d0-44c3-82c9-c5bbacf4271b-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:19:43 crc kubenswrapper[4733]: I1204 20:19:43.086231 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52qf2" Dec 04 20:19:43 crc kubenswrapper[4733]: I1204 20:19:43.135673 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-52qf2"] Dec 04 20:19:43 crc kubenswrapper[4733]: I1204 20:19:43.155535 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-52qf2"] Dec 04 20:19:44 crc kubenswrapper[4733]: I1204 20:19:44.354755 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" path="/var/lib/kubelet/pods/34d91348-64d0-44c3-82c9-c5bbacf4271b/volumes" Dec 04 20:20:02 crc kubenswrapper[4733]: I1204 20:20:02.192342 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 20:20:02 crc kubenswrapper[4733]: I1204 20:20:02.194558 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="54a0573d-d1fc-4355-a4e6-96a9db7a1a8f" containerName="adoption" containerID="cri-o://b108cc8a169835be119998e8931dce65f70fc8422eca3e347c388d28f51b1613" gracePeriod=30 Dec 04 20:20:15 crc kubenswrapper[4733]: I1204 20:20:15.362514 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:20:15 crc kubenswrapper[4733]: I1204 20:20:15.363285 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:20:30 crc kubenswrapper[4733]: I1204 20:20:30.100689 4733 scope.go:117] "RemoveContainer" containerID="7e6989849ccebd313015c5992fbf4660f5c31fc922f7af82b22d6052789a59fa" Dec 04 20:20:32 crc kubenswrapper[4733]: I1204 20:20:32.673154 4733 generic.go:334] "Generic (PLEG): container finished" podID="54a0573d-d1fc-4355-a4e6-96a9db7a1a8f" containerID="b108cc8a169835be119998e8931dce65f70fc8422eca3e347c388d28f51b1613" exitCode=137 Dec 04 20:20:32 crc kubenswrapper[4733]: I1204 20:20:32.673262 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f","Type":"ContainerDied","Data":"b108cc8a169835be119998e8931dce65f70fc8422eca3e347c388d28f51b1613"} Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.413837 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.437893 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkckc\" (UniqueName: \"kubernetes.io/projected/54a0573d-d1fc-4355-a4e6-96a9db7a1a8f-kube-api-access-vkckc\") pod \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") " Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.439362 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\") pod \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\" (UID: \"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f\") " Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.444037 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a0573d-d1fc-4355-a4e6-96a9db7a1a8f-kube-api-access-vkckc" (OuterVolumeSpecName: "kube-api-access-vkckc") pod "54a0573d-d1fc-4355-a4e6-96a9db7a1a8f" (UID: "54a0573d-d1fc-4355-a4e6-96a9db7a1a8f"). InnerVolumeSpecName "kube-api-access-vkckc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.458360 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346" (OuterVolumeSpecName: "mariadb-data") pod "54a0573d-d1fc-4355-a4e6-96a9db7a1a8f" (UID: "54a0573d-d1fc-4355-a4e6-96a9db7a1a8f"). InnerVolumeSpecName "pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.541649 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkckc\" (UniqueName: \"kubernetes.io/projected/54a0573d-d1fc-4355-a4e6-96a9db7a1a8f-kube-api-access-vkckc\") on node \"crc\" DevicePath \"\"" Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.541711 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\") on node \"crc\" " Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.568452 4733 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.568698 4733 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346") on node "crc" Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.650359 4733 reconciler_common.go:293] "Volume detached for volume \"pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6ecad577-2fde-46ba-97b7-ff8bf82e4346\") on node \"crc\" DevicePath \"\"" Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.687131 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"54a0573d-d1fc-4355-a4e6-96a9db7a1a8f","Type":"ContainerDied","Data":"af26d3bd6fcd8d1d31a7631394b5571a3ff799e96742e0ce6e5afa2164a0f752"} Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.687547 4733 scope.go:117] "RemoveContainer" containerID="b108cc8a169835be119998e8931dce65f70fc8422eca3e347c388d28f51b1613" Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.687276 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.727946 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 20:20:33 crc kubenswrapper[4733]: I1204 20:20:33.740257 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 20:20:34 crc kubenswrapper[4733]: I1204 20:20:34.349561 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a0573d-d1fc-4355-a4e6-96a9db7a1a8f" path="/var/lib/kubelet/pods/54a0573d-d1fc-4355-a4e6-96a9db7a1a8f/volumes" Dec 04 20:20:34 crc kubenswrapper[4733]: I1204 20:20:34.513322 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 20:20:34 crc kubenswrapper[4733]: I1204 20:20:34.513546 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="0c73247a-c966-401a-944d-31c43164b6b7" containerName="adoption" containerID="cri-o://b259aacc064a590c59a956ec0d0863a146ed74215a33276852ca9599d0d3e41b" gracePeriod=30 Dec 04 20:20:45 crc kubenswrapper[4733]: I1204 20:20:45.362106 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:20:45 crc kubenswrapper[4733]: I1204 20:20:45.362782 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.049465 4733 generic.go:334] "Generic (PLEG): container finished" podID="0c73247a-c966-401a-944d-31c43164b6b7" containerID="b259aacc064a590c59a956ec0d0863a146ed74215a33276852ca9599d0d3e41b" exitCode=137 Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.050114 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"0c73247a-c966-401a-944d-31c43164b6b7","Type":"ContainerDied","Data":"b259aacc064a590c59a956ec0d0863a146ed74215a33276852ca9599d0d3e41b"} Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.050150 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"0c73247a-c966-401a-944d-31c43164b6b7","Type":"ContainerDied","Data":"48f7f1a986182b71bc0535f509fca3d9f8cbb17c281cd0cd8179a54ab9e6bbe8"} Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.050167 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48f7f1a986182b71bc0535f509fca3d9f8cbb17c281cd0cd8179a54ab9e6bbe8" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.079422 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.179877 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtdpf\" (UniqueName: \"kubernetes.io/projected/0c73247a-c966-401a-944d-31c43164b6b7-kube-api-access-xtdpf\") pod \"0c73247a-c966-401a-944d-31c43164b6b7\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.180024 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0c73247a-c966-401a-944d-31c43164b6b7-ovn-data-cert\") pod \"0c73247a-c966-401a-944d-31c43164b6b7\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.180647 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\") pod \"0c73247a-c966-401a-944d-31c43164b6b7\" (UID: \"0c73247a-c966-401a-944d-31c43164b6b7\") " Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.185618 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c73247a-c966-401a-944d-31c43164b6b7-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "0c73247a-c966-401a-944d-31c43164b6b7" (UID: "0c73247a-c966-401a-944d-31c43164b6b7"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.191935 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c73247a-c966-401a-944d-31c43164b6b7-kube-api-access-xtdpf" (OuterVolumeSpecName: "kube-api-access-xtdpf") pod "0c73247a-c966-401a-944d-31c43164b6b7" (UID: "0c73247a-c966-401a-944d-31c43164b6b7"). InnerVolumeSpecName "kube-api-access-xtdpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.200225 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f" (OuterVolumeSpecName: "ovn-data") pod "0c73247a-c966-401a-944d-31c43164b6b7" (UID: "0c73247a-c966-401a-944d-31c43164b6b7"). InnerVolumeSpecName "pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.283454 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\") on node \"crc\" " Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.283488 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtdpf\" (UniqueName: \"kubernetes.io/projected/0c73247a-c966-401a-944d-31c43164b6b7-kube-api-access-xtdpf\") on node \"crc\" DevicePath \"\"" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.283499 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0c73247a-c966-401a-944d-31c43164b6b7-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.307276 4733 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.307459 4733 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f") on node "crc" Dec 04 20:21:05 crc kubenswrapper[4733]: I1204 20:21:05.385362 4733 reconciler_common.go:293] "Volume detached for volume \"pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d65dca1e-8249-4ebe-bf16-3d1ad911d07f\") on node \"crc\" DevicePath \"\"" Dec 04 20:21:06 crc kubenswrapper[4733]: I1204 20:21:06.072558 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 04 20:21:06 crc kubenswrapper[4733]: I1204 20:21:06.123602 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 20:21:06 crc kubenswrapper[4733]: I1204 20:21:06.145526 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 20:21:06 crc kubenswrapper[4733]: I1204 20:21:06.346767 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c73247a-c966-401a-944d-31c43164b6b7" path="/var/lib/kubelet/pods/0c73247a-c966-401a-944d-31c43164b6b7/volumes" Dec 04 20:21:15 crc kubenswrapper[4733]: I1204 20:21:15.361975 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:21:15 crc kubenswrapper[4733]: I1204 20:21:15.362882 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:21:15 crc kubenswrapper[4733]: I1204 20:21:15.362960 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 20:21:15 crc kubenswrapper[4733]: I1204 20:21:15.364341 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2ae00648d3f84b6e62e8b996f069c0f30a14ec5ac0a0886f49c65f077b2f9bda"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 20:21:15 crc kubenswrapper[4733]: I1204 20:21:15.364460 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://2ae00648d3f84b6e62e8b996f069c0f30a14ec5ac0a0886f49c65f077b2f9bda" gracePeriod=600 Dec 04 20:21:16 crc kubenswrapper[4733]: I1204 20:21:16.204754 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="2ae00648d3f84b6e62e8b996f069c0f30a14ec5ac0a0886f49c65f077b2f9bda" exitCode=0 Dec 04 20:21:16 crc kubenswrapper[4733]: I1204 20:21:16.204967 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"2ae00648d3f84b6e62e8b996f069c0f30a14ec5ac0a0886f49c65f077b2f9bda"} Dec 04 20:21:16 crc kubenswrapper[4733]: I1204 20:21:16.205458 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce"} Dec 04 20:21:16 crc kubenswrapper[4733]: I1204 20:21:16.205490 4733 scope.go:117] "RemoveContainer" containerID="db931af659084e1aaa786f8d3d321c3efd2f265811726c8b3ab51d6d68197039" Dec 04 20:21:30 crc kubenswrapper[4733]: I1204 20:21:30.156817 4733 scope.go:117] "RemoveContainer" containerID="b259aacc064a590c59a956ec0d0863a146ed74215a33276852ca9599d0d3e41b" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.593397 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n24qc/must-gather-phwjg"] Dec 04 20:22:12 crc kubenswrapper[4733]: E1204 20:22:12.594337 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a0573d-d1fc-4355-a4e6-96a9db7a1a8f" containerName="adoption" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.594349 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a0573d-d1fc-4355-a4e6-96a9db7a1a8f" containerName="adoption" Dec 04 20:22:12 crc kubenswrapper[4733]: E1204 20:22:12.594363 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerName="registry-server" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.594369 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerName="registry-server" Dec 04 20:22:12 crc kubenswrapper[4733]: E1204 20:22:12.594389 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c73247a-c966-401a-944d-31c43164b6b7" containerName="adoption" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.594396 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c73247a-c966-401a-944d-31c43164b6b7" containerName="adoption" Dec 04 20:22:12 crc kubenswrapper[4733]: E1204 20:22:12.594415 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerName="extract-content" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.594421 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerName="extract-content" Dec 04 20:22:12 crc kubenswrapper[4733]: E1204 20:22:12.594445 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerName="extract-utilities" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.594451 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerName="extract-utilities" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.594637 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a0573d-d1fc-4355-a4e6-96a9db7a1a8f" containerName="adoption" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.594650 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c73247a-c966-401a-944d-31c43164b6b7" containerName="adoption" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.594678 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="34d91348-64d0-44c3-82c9-c5bbacf4271b" containerName="registry-server" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.595884 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.597892 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n24qc"/"openshift-service-ca.crt" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.598205 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-n24qc"/"default-dockercfg-ls6t9" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.598333 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n24qc"/"kube-root-ca.crt" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.613639 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n24qc/must-gather-phwjg"] Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.772548 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc2wf\" (UniqueName: \"kubernetes.io/projected/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-kube-api-access-hc2wf\") pod \"must-gather-phwjg\" (UID: \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\") " pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.772660 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-must-gather-output\") pod \"must-gather-phwjg\" (UID: \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\") " pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.874884 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc2wf\" (UniqueName: \"kubernetes.io/projected/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-kube-api-access-hc2wf\") pod \"must-gather-phwjg\" (UID: \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\") " pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.875028 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-must-gather-output\") pod \"must-gather-phwjg\" (UID: \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\") " pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:22:12 crc kubenswrapper[4733]: I1204 20:22:12.875559 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-must-gather-output\") pod \"must-gather-phwjg\" (UID: \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\") " pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:22:13 crc kubenswrapper[4733]: I1204 20:22:13.443539 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc2wf\" (UniqueName: \"kubernetes.io/projected/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-kube-api-access-hc2wf\") pod \"must-gather-phwjg\" (UID: \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\") " pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:22:13 crc kubenswrapper[4733]: I1204 20:22:13.513759 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:22:13 crc kubenswrapper[4733]: I1204 20:22:13.998365 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n24qc/must-gather-phwjg"] Dec 04 20:22:14 crc kubenswrapper[4733]: I1204 20:22:14.170563 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/must-gather-phwjg" event={"ID":"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0","Type":"ContainerStarted","Data":"b78919a47feca08d4a8be7b8129db1438b8735b2ae0fceb159634df32aa17e47"} Dec 04 20:22:19 crc kubenswrapper[4733]: I1204 20:22:19.217861 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/must-gather-phwjg" event={"ID":"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0","Type":"ContainerStarted","Data":"0496e1082cbe5bd8b0c9759cadf39d18f3a1d2dd09604bb8a9b969589e3c3486"} Dec 04 20:22:19 crc kubenswrapper[4733]: I1204 20:22:19.218487 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/must-gather-phwjg" event={"ID":"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0","Type":"ContainerStarted","Data":"5926f52b4eb9210fadba47185d2905a28d1a3ef594789f628b6b7275ad423935"} Dec 04 20:22:19 crc kubenswrapper[4733]: I1204 20:22:19.243002 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n24qc/must-gather-phwjg" podStartSLOduration=3.017768086 podStartE2EDuration="7.242981079s" podCreationTimestamp="2025-12-04 20:22:12 +0000 UTC" firstStartedPulling="2025-12-04 20:22:14.012297119 +0000 UTC m=+9795.967658165" lastFinishedPulling="2025-12-04 20:22:18.237510112 +0000 UTC m=+9800.192871158" observedRunningTime="2025-12-04 20:22:19.230300285 +0000 UTC m=+9801.185661331" watchObservedRunningTime="2025-12-04 20:22:19.242981079 +0000 UTC m=+9801.198342135" Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.522807 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n24qc/crc-debug-f5wxt"] Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.524829 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.694359 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lrh2\" (UniqueName: \"kubernetes.io/projected/373dd85a-2aa0-4188-baae-09f82aa0f24d-kube-api-access-8lrh2\") pod \"crc-debug-f5wxt\" (UID: \"373dd85a-2aa0-4188-baae-09f82aa0f24d\") " pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.694445 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/373dd85a-2aa0-4188-baae-09f82aa0f24d-host\") pod \"crc-debug-f5wxt\" (UID: \"373dd85a-2aa0-4188-baae-09f82aa0f24d\") " pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.797084 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/373dd85a-2aa0-4188-baae-09f82aa0f24d-host\") pod \"crc-debug-f5wxt\" (UID: \"373dd85a-2aa0-4188-baae-09f82aa0f24d\") " pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.797262 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/373dd85a-2aa0-4188-baae-09f82aa0f24d-host\") pod \"crc-debug-f5wxt\" (UID: \"373dd85a-2aa0-4188-baae-09f82aa0f24d\") " pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.797839 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lrh2\" (UniqueName: \"kubernetes.io/projected/373dd85a-2aa0-4188-baae-09f82aa0f24d-kube-api-access-8lrh2\") pod \"crc-debug-f5wxt\" (UID: \"373dd85a-2aa0-4188-baae-09f82aa0f24d\") " pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.828989 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lrh2\" (UniqueName: \"kubernetes.io/projected/373dd85a-2aa0-4188-baae-09f82aa0f24d-kube-api-access-8lrh2\") pod \"crc-debug-f5wxt\" (UID: \"373dd85a-2aa0-4188-baae-09f82aa0f24d\") " pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:22:22 crc kubenswrapper[4733]: I1204 20:22:22.849089 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:22:22 crc kubenswrapper[4733]: W1204 20:22:22.895359 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod373dd85a_2aa0_4188_baae_09f82aa0f24d.slice/crio-9b0ff06152c29962b1c2c8dbd8cc44788350a268293e81b8c36c45a924fb7461 WatchSource:0}: Error finding container 9b0ff06152c29962b1c2c8dbd8cc44788350a268293e81b8c36c45a924fb7461: Status 404 returned error can't find the container with id 9b0ff06152c29962b1c2c8dbd8cc44788350a268293e81b8c36c45a924fb7461 Dec 04 20:22:23 crc kubenswrapper[4733]: I1204 20:22:23.273765 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/crc-debug-f5wxt" event={"ID":"373dd85a-2aa0-4188-baae-09f82aa0f24d","Type":"ContainerStarted","Data":"9b0ff06152c29962b1c2c8dbd8cc44788350a268293e81b8c36c45a924fb7461"} Dec 04 20:22:35 crc kubenswrapper[4733]: I1204 20:22:35.415140 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/crc-debug-f5wxt" event={"ID":"373dd85a-2aa0-4188-baae-09f82aa0f24d","Type":"ContainerStarted","Data":"2010b5b375dbd0132eb5a265d4dcceae05599b30cfe29333b638bfd535719c55"} Dec 04 20:22:35 crc kubenswrapper[4733]: I1204 20:22:35.432357 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n24qc/crc-debug-f5wxt" podStartSLOduration=1.511535216 podStartE2EDuration="13.432339751s" podCreationTimestamp="2025-12-04 20:22:22 +0000 UTC" firstStartedPulling="2025-12-04 20:22:22.89777856 +0000 UTC m=+9804.853139616" lastFinishedPulling="2025-12-04 20:22:34.818583105 +0000 UTC m=+9816.773944151" observedRunningTime="2025-12-04 20:22:35.428113186 +0000 UTC m=+9817.383474282" watchObservedRunningTime="2025-12-04 20:22:35.432339751 +0000 UTC m=+9817.387700797" Dec 04 20:22:59 crc kubenswrapper[4733]: I1204 20:22:59.647899 4733 generic.go:334] "Generic (PLEG): container finished" podID="373dd85a-2aa0-4188-baae-09f82aa0f24d" containerID="2010b5b375dbd0132eb5a265d4dcceae05599b30cfe29333b638bfd535719c55" exitCode=0 Dec 04 20:22:59 crc kubenswrapper[4733]: I1204 20:22:59.647966 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/crc-debug-f5wxt" event={"ID":"373dd85a-2aa0-4188-baae-09f82aa0f24d","Type":"ContainerDied","Data":"2010b5b375dbd0132eb5a265d4dcceae05599b30cfe29333b638bfd535719c55"} Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.768271 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.795739 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/373dd85a-2aa0-4188-baae-09f82aa0f24d-host\") pod \"373dd85a-2aa0-4188-baae-09f82aa0f24d\" (UID: \"373dd85a-2aa0-4188-baae-09f82aa0f24d\") " Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.795853 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lrh2\" (UniqueName: \"kubernetes.io/projected/373dd85a-2aa0-4188-baae-09f82aa0f24d-kube-api-access-8lrh2\") pod \"373dd85a-2aa0-4188-baae-09f82aa0f24d\" (UID: \"373dd85a-2aa0-4188-baae-09f82aa0f24d\") " Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.795854 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/373dd85a-2aa0-4188-baae-09f82aa0f24d-host" (OuterVolumeSpecName: "host") pod "373dd85a-2aa0-4188-baae-09f82aa0f24d" (UID: "373dd85a-2aa0-4188-baae-09f82aa0f24d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.796280 4733 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/373dd85a-2aa0-4188-baae-09f82aa0f24d-host\") on node \"crc\" DevicePath \"\"" Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.814726 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373dd85a-2aa0-4188-baae-09f82aa0f24d-kube-api-access-8lrh2" (OuterVolumeSpecName: "kube-api-access-8lrh2") pod "373dd85a-2aa0-4188-baae-09f82aa0f24d" (UID: "373dd85a-2aa0-4188-baae-09f82aa0f24d"). InnerVolumeSpecName "kube-api-access-8lrh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.826404 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n24qc/crc-debug-f5wxt"] Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.837397 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n24qc/crc-debug-f5wxt"] Dec 04 20:23:00 crc kubenswrapper[4733]: I1204 20:23:00.897806 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lrh2\" (UniqueName: \"kubernetes.io/projected/373dd85a-2aa0-4188-baae-09f82aa0f24d-kube-api-access-8lrh2\") on node \"crc\" DevicePath \"\"" Dec 04 20:23:01 crc kubenswrapper[4733]: I1204 20:23:01.706422 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b0ff06152c29962b1c2c8dbd8cc44788350a268293e81b8c36c45a924fb7461" Dec 04 20:23:01 crc kubenswrapper[4733]: I1204 20:23:01.707076 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/crc-debug-f5wxt" Dec 04 20:23:01 crc kubenswrapper[4733]: I1204 20:23:01.997061 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n24qc/crc-debug-k2n84"] Dec 04 20:23:01 crc kubenswrapper[4733]: E1204 20:23:01.997631 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373dd85a-2aa0-4188-baae-09f82aa0f24d" containerName="container-00" Dec 04 20:23:01 crc kubenswrapper[4733]: I1204 20:23:01.997649 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="373dd85a-2aa0-4188-baae-09f82aa0f24d" containerName="container-00" Dec 04 20:23:01 crc kubenswrapper[4733]: I1204 20:23:01.997896 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="373dd85a-2aa0-4188-baae-09f82aa0f24d" containerName="container-00" Dec 04 20:23:01 crc kubenswrapper[4733]: I1204 20:23:01.998731 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.125439 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-host\") pod \"crc-debug-k2n84\" (UID: \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\") " pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.125558 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccvpz\" (UniqueName: \"kubernetes.io/projected/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-kube-api-access-ccvpz\") pod \"crc-debug-k2n84\" (UID: \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\") " pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.227926 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-host\") pod \"crc-debug-k2n84\" (UID: \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\") " pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.228308 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccvpz\" (UniqueName: \"kubernetes.io/projected/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-kube-api-access-ccvpz\") pod \"crc-debug-k2n84\" (UID: \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\") " pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.228103 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-host\") pod \"crc-debug-k2n84\" (UID: \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\") " pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.258619 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccvpz\" (UniqueName: \"kubernetes.io/projected/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-kube-api-access-ccvpz\") pod \"crc-debug-k2n84\" (UID: \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\") " pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.314664 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.351029 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="373dd85a-2aa0-4188-baae-09f82aa0f24d" path="/var/lib/kubelet/pods/373dd85a-2aa0-4188-baae-09f82aa0f24d/volumes" Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.717643 4733 generic.go:334] "Generic (PLEG): container finished" podID="cd24d43e-fd75-45d0-97de-6cce8d3f4d4d" containerID="5648c723471aacae32b9d52987d45ee60ee7955f62f7e5456f0046d99b0a6b70" exitCode=1 Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.717743 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/crc-debug-k2n84" event={"ID":"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d","Type":"ContainerDied","Data":"5648c723471aacae32b9d52987d45ee60ee7955f62f7e5456f0046d99b0a6b70"} Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.717969 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/crc-debug-k2n84" event={"ID":"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d","Type":"ContainerStarted","Data":"0d2ab25a416a406eb73e6e1eeabf730bfc158a6c62584fb61e999b2d60812cd4"} Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.759370 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n24qc/crc-debug-k2n84"] Dec 04 20:23:02 crc kubenswrapper[4733]: I1204 20:23:02.768266 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n24qc/crc-debug-k2n84"] Dec 04 20:23:03 crc kubenswrapper[4733]: I1204 20:23:03.841229 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:03 crc kubenswrapper[4733]: I1204 20:23:03.965178 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-host\") pod \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\" (UID: \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\") " Dec 04 20:23:03 crc kubenswrapper[4733]: I1204 20:23:03.965219 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccvpz\" (UniqueName: \"kubernetes.io/projected/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-kube-api-access-ccvpz\") pod \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\" (UID: \"cd24d43e-fd75-45d0-97de-6cce8d3f4d4d\") " Dec 04 20:23:03 crc kubenswrapper[4733]: I1204 20:23:03.966531 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-host" (OuterVolumeSpecName: "host") pod "cd24d43e-fd75-45d0-97de-6cce8d3f4d4d" (UID: "cd24d43e-fd75-45d0-97de-6cce8d3f4d4d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 20:23:03 crc kubenswrapper[4733]: I1204 20:23:03.974093 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-kube-api-access-ccvpz" (OuterVolumeSpecName: "kube-api-access-ccvpz") pod "cd24d43e-fd75-45d0-97de-6cce8d3f4d4d" (UID: "cd24d43e-fd75-45d0-97de-6cce8d3f4d4d"). InnerVolumeSpecName "kube-api-access-ccvpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:23:04 crc kubenswrapper[4733]: I1204 20:23:04.068555 4733 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-host\") on node \"crc\" DevicePath \"\"" Dec 04 20:23:04 crc kubenswrapper[4733]: I1204 20:23:04.068603 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccvpz\" (UniqueName: \"kubernetes.io/projected/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d-kube-api-access-ccvpz\") on node \"crc\" DevicePath \"\"" Dec 04 20:23:04 crc kubenswrapper[4733]: I1204 20:23:04.370380 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd24d43e-fd75-45d0-97de-6cce8d3f4d4d" path="/var/lib/kubelet/pods/cd24d43e-fd75-45d0-97de-6cce8d3f4d4d/volumes" Dec 04 20:23:04 crc kubenswrapper[4733]: I1204 20:23:04.738239 4733 scope.go:117] "RemoveContainer" containerID="5648c723471aacae32b9d52987d45ee60ee7955f62f7e5456f0046d99b0a6b70" Dec 04 20:23:04 crc kubenswrapper[4733]: I1204 20:23:04.738293 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/crc-debug-k2n84" Dec 04 20:23:15 crc kubenswrapper[4733]: I1204 20:23:15.362196 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:23:15 crc kubenswrapper[4733]: I1204 20:23:15.363594 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:23:45 crc kubenswrapper[4733]: I1204 20:23:45.361583 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:23:45 crc kubenswrapper[4733]: I1204 20:23:45.362300 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.362648 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.363297 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.363352 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.364180 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.364256 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" gracePeriod=600 Dec 04 20:24:15 crc kubenswrapper[4733]: E1204 20:24:15.486173 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.550895 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" exitCode=0 Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.550941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce"} Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.550980 4733 scope.go:117] "RemoveContainer" containerID="2ae00648d3f84b6e62e8b996f069c0f30a14ec5ac0a0886f49c65f077b2f9bda" Dec 04 20:24:15 crc kubenswrapper[4733]: I1204 20:24:15.551589 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:24:15 crc kubenswrapper[4733]: E1204 20:24:15.551871 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:24:26 crc kubenswrapper[4733]: I1204 20:24:26.336206 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:24:26 crc kubenswrapper[4733]: E1204 20:24:26.337547 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:24:37 crc kubenswrapper[4733]: I1204 20:24:37.962200 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7bz4t"] Dec 04 20:24:37 crc kubenswrapper[4733]: E1204 20:24:37.963338 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd24d43e-fd75-45d0-97de-6cce8d3f4d4d" containerName="container-00" Dec 04 20:24:37 crc kubenswrapper[4733]: I1204 20:24:37.963356 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd24d43e-fd75-45d0-97de-6cce8d3f4d4d" containerName="container-00" Dec 04 20:24:37 crc kubenswrapper[4733]: I1204 20:24:37.963658 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd24d43e-fd75-45d0-97de-6cce8d3f4d4d" containerName="container-00" Dec 04 20:24:37 crc kubenswrapper[4733]: I1204 20:24:37.966141 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:37 crc kubenswrapper[4733]: I1204 20:24:37.972021 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7bz4t"] Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.029862 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-utilities\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.029971 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-catalog-content\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.030012 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l9qk\" (UniqueName: \"kubernetes.io/projected/b604bde3-fe9e-43a9-9c02-fd645889b821-kube-api-access-4l9qk\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.131903 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-catalog-content\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.131967 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l9qk\" (UniqueName: \"kubernetes.io/projected/b604bde3-fe9e-43a9-9c02-fd645889b821-kube-api-access-4l9qk\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.132083 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-utilities\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.132497 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-catalog-content\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.132551 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-utilities\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.149317 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l9qk\" (UniqueName: \"kubernetes.io/projected/b604bde3-fe9e-43a9-9c02-fd645889b821-kube-api-access-4l9qk\") pod \"certified-operators-7bz4t\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.289771 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:38 crc kubenswrapper[4733]: I1204 20:24:38.890898 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7bz4t"] Dec 04 20:24:39 crc kubenswrapper[4733]: I1204 20:24:39.842096 4733 generic.go:334] "Generic (PLEG): container finished" podID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerID="e58fdf3e112c8712c75cfa37903fefa157ca1c56aa9865be50e30424fa5713b2" exitCode=0 Dec 04 20:24:39 crc kubenswrapper[4733]: I1204 20:24:39.842157 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bz4t" event={"ID":"b604bde3-fe9e-43a9-9c02-fd645889b821","Type":"ContainerDied","Data":"e58fdf3e112c8712c75cfa37903fefa157ca1c56aa9865be50e30424fa5713b2"} Dec 04 20:24:39 crc kubenswrapper[4733]: I1204 20:24:39.842589 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bz4t" event={"ID":"b604bde3-fe9e-43a9-9c02-fd645889b821","Type":"ContainerStarted","Data":"7d4d3ac60dbed13b6fa2843ef7f6e2b3bb17b7d72101bf887db334bb67d01509"} Dec 04 20:24:39 crc kubenswrapper[4733]: I1204 20:24:39.844859 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 20:24:40 crc kubenswrapper[4733]: I1204 20:24:40.854453 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bz4t" event={"ID":"b604bde3-fe9e-43a9-9c02-fd645889b821","Type":"ContainerStarted","Data":"8b515b5206cbc162275d4b903384638e803f75a652e4f73387fa6369c3fd10d4"} Dec 04 20:24:41 crc kubenswrapper[4733]: I1204 20:24:41.335719 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:24:41 crc kubenswrapper[4733]: E1204 20:24:41.336161 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:24:41 crc kubenswrapper[4733]: I1204 20:24:41.868170 4733 generic.go:334] "Generic (PLEG): container finished" podID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerID="8b515b5206cbc162275d4b903384638e803f75a652e4f73387fa6369c3fd10d4" exitCode=0 Dec 04 20:24:41 crc kubenswrapper[4733]: I1204 20:24:41.868206 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bz4t" event={"ID":"b604bde3-fe9e-43a9-9c02-fd645889b821","Type":"ContainerDied","Data":"8b515b5206cbc162275d4b903384638e803f75a652e4f73387fa6369c3fd10d4"} Dec 04 20:24:42 crc kubenswrapper[4733]: I1204 20:24:42.884499 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bz4t" event={"ID":"b604bde3-fe9e-43a9-9c02-fd645889b821","Type":"ContainerStarted","Data":"06582b122a58e006b681c03d765940d8c7bfe210941badf944c4310b2340dc22"} Dec 04 20:24:42 crc kubenswrapper[4733]: I1204 20:24:42.917520 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7bz4t" podStartSLOduration=3.3273679720000002 podStartE2EDuration="5.917499573s" podCreationTimestamp="2025-12-04 20:24:37 +0000 UTC" firstStartedPulling="2025-12-04 20:24:39.844575283 +0000 UTC m=+9941.799936329" lastFinishedPulling="2025-12-04 20:24:42.434706884 +0000 UTC m=+9944.390067930" observedRunningTime="2025-12-04 20:24:42.90375847 +0000 UTC m=+9944.859119546" watchObservedRunningTime="2025-12-04 20:24:42.917499573 +0000 UTC m=+9944.872860639" Dec 04 20:24:48 crc kubenswrapper[4733]: I1204 20:24:48.290442 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:48 crc kubenswrapper[4733]: I1204 20:24:48.291167 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:48 crc kubenswrapper[4733]: I1204 20:24:48.401928 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:49 crc kubenswrapper[4733]: I1204 20:24:49.030618 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:49 crc kubenswrapper[4733]: I1204 20:24:49.097035 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7bz4t"] Dec 04 20:24:50 crc kubenswrapper[4733]: I1204 20:24:50.971175 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7bz4t" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerName="registry-server" containerID="cri-o://06582b122a58e006b681c03d765940d8c7bfe210941badf944c4310b2340dc22" gracePeriod=2 Dec 04 20:24:51 crc kubenswrapper[4733]: I1204 20:24:51.989997 4733 generic.go:334] "Generic (PLEG): container finished" podID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerID="06582b122a58e006b681c03d765940d8c7bfe210941badf944c4310b2340dc22" exitCode=0 Dec 04 20:24:51 crc kubenswrapper[4733]: I1204 20:24:51.990077 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bz4t" event={"ID":"b604bde3-fe9e-43a9-9c02-fd645889b821","Type":"ContainerDied","Data":"06582b122a58e006b681c03d765940d8c7bfe210941badf944c4310b2340dc22"} Dec 04 20:24:51 crc kubenswrapper[4733]: I1204 20:24:51.990384 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bz4t" event={"ID":"b604bde3-fe9e-43a9-9c02-fd645889b821","Type":"ContainerDied","Data":"7d4d3ac60dbed13b6fa2843ef7f6e2b3bb17b7d72101bf887db334bb67d01509"} Dec 04 20:24:51 crc kubenswrapper[4733]: I1204 20:24:51.990401 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d4d3ac60dbed13b6fa2843ef7f6e2b3bb17b7d72101bf887db334bb67d01509" Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.097357 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.166951 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-catalog-content\") pod \"b604bde3-fe9e-43a9-9c02-fd645889b821\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.167060 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l9qk\" (UniqueName: \"kubernetes.io/projected/b604bde3-fe9e-43a9-9c02-fd645889b821-kube-api-access-4l9qk\") pod \"b604bde3-fe9e-43a9-9c02-fd645889b821\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.167133 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-utilities\") pod \"b604bde3-fe9e-43a9-9c02-fd645889b821\" (UID: \"b604bde3-fe9e-43a9-9c02-fd645889b821\") " Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.168068 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-utilities" (OuterVolumeSpecName: "utilities") pod "b604bde3-fe9e-43a9-9c02-fd645889b821" (UID: "b604bde3-fe9e-43a9-9c02-fd645889b821"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.187327 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b604bde3-fe9e-43a9-9c02-fd645889b821-kube-api-access-4l9qk" (OuterVolumeSpecName: "kube-api-access-4l9qk") pod "b604bde3-fe9e-43a9-9c02-fd645889b821" (UID: "b604bde3-fe9e-43a9-9c02-fd645889b821"). InnerVolumeSpecName "kube-api-access-4l9qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.223142 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b604bde3-fe9e-43a9-9c02-fd645889b821" (UID: "b604bde3-fe9e-43a9-9c02-fd645889b821"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.270103 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.270136 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l9qk\" (UniqueName: \"kubernetes.io/projected/b604bde3-fe9e-43a9-9c02-fd645889b821-kube-api-access-4l9qk\") on node \"crc\" DevicePath \"\"" Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.270149 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b604bde3-fe9e-43a9-9c02-fd645889b821-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:24:52 crc kubenswrapper[4733]: I1204 20:24:52.337085 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:24:52 crc kubenswrapper[4733]: E1204 20:24:52.337343 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:24:53 crc kubenswrapper[4733]: I1204 20:24:53.000063 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bz4t" Dec 04 20:24:53 crc kubenswrapper[4733]: I1204 20:24:53.024558 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7bz4t"] Dec 04 20:24:53 crc kubenswrapper[4733]: I1204 20:24:53.034235 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7bz4t"] Dec 04 20:24:54 crc kubenswrapper[4733]: I1204 20:24:54.350653 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" path="/var/lib/kubelet/pods/b604bde3-fe9e-43a9-9c02-fd645889b821/volumes" Dec 04 20:25:07 crc kubenswrapper[4733]: I1204 20:25:07.335725 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:25:07 crc kubenswrapper[4733]: E1204 20:25:07.337054 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:25:21 crc kubenswrapper[4733]: I1204 20:25:21.335756 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:25:21 crc kubenswrapper[4733]: E1204 20:25:21.336757 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:25:30 crc kubenswrapper[4733]: I1204 20:25:30.544179 4733 scope.go:117] "RemoveContainer" containerID="f70a266c02b8e3fa0e77f5393bf0ef36f89ba699505cd0104b3c22b791eff081" Dec 04 20:25:35 crc kubenswrapper[4733]: I1204 20:25:35.335949 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:25:35 crc kubenswrapper[4733]: E1204 20:25:35.336567 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:25:48 crc kubenswrapper[4733]: I1204 20:25:48.342739 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:25:48 crc kubenswrapper[4733]: E1204 20:25:48.343959 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:26:00 crc kubenswrapper[4733]: I1204 20:26:00.335898 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:26:00 crc kubenswrapper[4733]: E1204 20:26:00.336751 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:26:14 crc kubenswrapper[4733]: I1204 20:26:14.335569 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:26:14 crc kubenswrapper[4733]: E1204 20:26:14.336688 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:26:25 crc kubenswrapper[4733]: I1204 20:26:25.335731 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:26:25 crc kubenswrapper[4733]: E1204 20:26:25.336725 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:26:30 crc kubenswrapper[4733]: I1204 20:26:30.609298 4733 scope.go:117] "RemoveContainer" containerID="524b989bae1bd4467a8b331e0b0b0a5fb8f4bce3cb713f5c3cfae086faa35475" Dec 04 20:26:30 crc kubenswrapper[4733]: I1204 20:26:30.639253 4733 scope.go:117] "RemoveContainer" containerID="b825261551734f7768c0ff3465fa00437c240137810eee029b584beb03ebbfb2" Dec 04 20:26:40 crc kubenswrapper[4733]: I1204 20:26:40.335632 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:26:40 crc kubenswrapper[4733]: E1204 20:26:40.336433 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:26:51 crc kubenswrapper[4733]: I1204 20:26:51.336778 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:26:51 crc kubenswrapper[4733]: E1204 20:26:51.337902 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:27:03 crc kubenswrapper[4733]: I1204 20:27:03.389823 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:27:03 crc kubenswrapper[4733]: E1204 20:27:03.395851 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:27:17 crc kubenswrapper[4733]: I1204 20:27:17.336482 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:27:17 crc kubenswrapper[4733]: E1204 20:27:17.337477 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:27:32 crc kubenswrapper[4733]: I1204 20:27:32.343138 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:27:32 crc kubenswrapper[4733]: E1204 20:27:32.343860 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:27:46 crc kubenswrapper[4733]: I1204 20:27:46.337034 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:27:46 crc kubenswrapper[4733]: E1204 20:27:46.338681 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:27:58 crc kubenswrapper[4733]: I1204 20:27:58.352318 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:27:58 crc kubenswrapper[4733]: E1204 20:27:58.353416 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:28:09 crc kubenswrapper[4733]: I1204 20:28:09.336465 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:28:09 crc kubenswrapper[4733]: E1204 20:28:09.340034 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:28:22 crc kubenswrapper[4733]: I1204 20:28:22.336597 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:28:22 crc kubenswrapper[4733]: E1204 20:28:22.337740 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:28:37 crc kubenswrapper[4733]: I1204 20:28:37.336471 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:28:37 crc kubenswrapper[4733]: E1204 20:28:37.337332 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:28:51 crc kubenswrapper[4733]: I1204 20:28:51.335447 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:28:51 crc kubenswrapper[4733]: E1204 20:28:51.336465 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:29:04 crc kubenswrapper[4733]: I1204 20:29:04.335636 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:29:04 crc kubenswrapper[4733]: E1204 20:29:04.336424 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:29:19 crc kubenswrapper[4733]: I1204 20:29:19.335762 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:29:19 crc kubenswrapper[4733]: I1204 20:29:19.565481 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_308b5993-05c7-4915-b744-8ac8797da024/init-config-reloader/0.log" Dec 04 20:29:19 crc kubenswrapper[4733]: I1204 20:29:19.695481 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_308b5993-05c7-4915-b744-8ac8797da024/init-config-reloader/0.log" Dec 04 20:29:19 crc kubenswrapper[4733]: I1204 20:29:19.781337 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_308b5993-05c7-4915-b744-8ac8797da024/alertmanager/0.log" Dec 04 20:29:19 crc kubenswrapper[4733]: I1204 20:29:19.814633 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_308b5993-05c7-4915-b744-8ac8797da024/config-reloader/0.log" Dec 04 20:29:19 crc kubenswrapper[4733]: I1204 20:29:19.903029 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_fc95841e-bba4-46fa-ae27-1b956e0d3a0c/aodh-api/0.log" Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.030484 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_fc95841e-bba4-46fa-ae27-1b956e0d3a0c/aodh-evaluator/0.log" Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.060683 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_fc95841e-bba4-46fa-ae27-1b956e0d3a0c/aodh-listener/0.log" Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.073000 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"fb18a7d2636425cdb65bf7d0042b43bec6a6390071894b7c5044793db67a3f33"} Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.185736 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_fc95841e-bba4-46fa-ae27-1b956e0d3a0c/aodh-notifier/0.log" Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.378957 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55cd5f5578-jr2bf_ae040e65-3459-4cdc-a35b-e2a47f6bcce3/barbican-api/0.log" Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.440994 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55cd5f5578-jr2bf_ae040e65-3459-4cdc-a35b-e2a47f6bcce3/barbican-api-log/0.log" Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.591044 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5bc7fbd458-mz9t9_a3871247-7c7c-4ccc-8c58-673cb82b7e97/barbican-keystone-listener/0.log" Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.663520 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5bc7fbd458-mz9t9_a3871247-7c7c-4ccc-8c58-673cb82b7e97/barbican-keystone-listener-log/0.log" Dec 04 20:29:20 crc kubenswrapper[4733]: I1204 20:29:20.788291 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68d4f54797-jj274_66a0bca5-dfe9-40b9-9d30-da5c46f887a2/barbican-worker/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.249418 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68d4f54797-jj274_66a0bca5-dfe9-40b9-9d30-da5c46f887a2/barbican-worker-log/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.257451 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_46d793fc-05ac-4707-9457-e11c52ab66a9/ceilometer-central-agent/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.292609 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-7q5rg_f85c314e-3306-4fa3-a717-9f31c1d4817f/bootstrap-openstack-openstack-cell1/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.437863 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_46d793fc-05ac-4707-9457-e11c52ab66a9/ceilometer-notification-agent/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.466415 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_46d793fc-05ac-4707-9457-e11c52ab66a9/sg-core/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.467667 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_46d793fc-05ac-4707-9457-e11c52ab66a9/proxy-httpd/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.659306 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-mpcsl_2bbe511a-0c11-477f-a274-81556cc13c97/ceph-client-openstack-openstack-cell1/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.762622 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e3c816c8-8307-4a11-a4ba-33615272c2f0/cinder-api/0.log" Dec 04 20:29:21 crc kubenswrapper[4733]: I1204 20:29:21.798059 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e3c816c8-8307-4a11-a4ba-33615272c2f0/cinder-api-log/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.076988 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_639c8ab0-c11a-4cf1-b944-9caa6b91604e/probe/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.146270 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_639c8ab0-c11a-4cf1-b944-9caa6b91604e/cinder-backup/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.206854 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f32fb6b2-626f-4319-92cb-3a129ec9ec7d/cinder-scheduler/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.352263 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f32fb6b2-626f-4319-92cb-3a129ec9ec7d/probe/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.427280 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_bb94a69c-3457-498f-a83a-43e375f89dfd/cinder-volume/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.463702 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_bb94a69c-3457-498f-a83a-43e375f89dfd/probe/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.654579 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-9v588_1b7b4789-500e-40d1-bddd-332ab22536e9/configure-network-openstack-openstack-cell1/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.673514 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-l82wt_27641fdb-1ce8-44d2-b0f6-f0b8cc303338/configure-os-openstack-openstack-cell1/0.log" Dec 04 20:29:22 crc kubenswrapper[4733]: I1204 20:29:22.870378 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d959b8c7-kxj9z_fbe1ced6-e57e-4add-8c40-a56fd21e055b/init/0.log" Dec 04 20:29:23 crc kubenswrapper[4733]: I1204 20:29:23.066746 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d959b8c7-kxj9z_fbe1ced6-e57e-4add-8c40-a56fd21e055b/init/0.log" Dec 04 20:29:23 crc kubenswrapper[4733]: I1204 20:29:23.133290 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d959b8c7-kxj9z_fbe1ced6-e57e-4add-8c40-a56fd21e055b/dnsmasq-dns/0.log" Dec 04 20:29:23 crc kubenswrapper[4733]: I1204 20:29:23.627392 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9455ff4a-328e-4b58-bd3e-a3e7b11c7737/glance-httpd/0.log" Dec 04 20:29:23 crc kubenswrapper[4733]: I1204 20:29:23.648415 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-s9lm6_b149b3e0-ab05-4b47-ac2d-d1d9e2d98b79/download-cache-openstack-openstack-cell1/0.log" Dec 04 20:29:23 crc kubenswrapper[4733]: I1204 20:29:23.742203 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9455ff4a-328e-4b58-bd3e-a3e7b11c7737/glance-log/0.log" Dec 04 20:29:23 crc kubenswrapper[4733]: I1204 20:29:23.875599 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b95d20bf-4ad1-438a-80e2-34fbac20bb7e/glance-log/0.log" Dec 04 20:29:23 crc kubenswrapper[4733]: I1204 20:29:23.888290 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b95d20bf-4ad1-438a-80e2-34fbac20bb7e/glance-httpd/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.048054 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-686759f4c9-dw5h2_865712a6-df7e-486a-9cae-bd0a0de5556a/heat-api/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.230448 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-66c8bffccb-r24sl_37ac4820-fee5-490a-83be-48d01d292e8f/heat-cfnapi/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.320648 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-68ff895d44-vq5jt_808a8006-fe18-43ee-b546-4a6043eb1081/heat-engine/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.461926 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5884fc67b7-6bjb4_21cdbfda-59ac-4b40-ab39-8efde3ba9de7/horizon/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.520905 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5884fc67b7-6bjb4_21cdbfda-59ac-4b40-ab39-8efde3ba9de7/horizon-log/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.570364 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-7w94z_271ce371-da5c-40a8-9f8f-bc90fa1bee8d/install-certs-openstack-openstack-cell1/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.705876 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-fpctx_0899052c-379c-4fd4-869b-b49d488cc60b/install-os-openstack-openstack-cell1/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.900288 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29414641-j7k8n_e83b39f0-0ffc-46d2-83b8-fbccf10aeef8/keystone-cron/0.log" Dec 04 20:29:24 crc kubenswrapper[4733]: I1204 20:29:24.914348 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5f9c8c86c8-h296q_923bc92b-30a9-48c8-9228-46f7e257db9a/keystone-api/0.log" Dec 04 20:29:25 crc kubenswrapper[4733]: I1204 20:29:25.018412 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a2fce55f-036a-419c-b710-17eafaa65162/kube-state-metrics/0.log" Dec 04 20:29:25 crc kubenswrapper[4733]: I1204 20:29:25.581394 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-zk2ps_df93d53f-da19-405c-a337-bbce7af8fd33/libvirt-openstack-openstack-cell1/0.log" Dec 04 20:29:25 crc kubenswrapper[4733]: I1204 20:29:25.660021 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8b43f863-6942-4b7b-a4e5-9f6265204ad4/manila-api-log/0.log" Dec 04 20:29:25 crc kubenswrapper[4733]: I1204 20:29:25.718264 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8b43f863-6942-4b7b-a4e5-9f6265204ad4/manila-api/0.log" Dec 04 20:29:25 crc kubenswrapper[4733]: I1204 20:29:25.877626 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ccbe2eef-52a5-41b9-9000-bb0376d74bdc/probe/0.log" Dec 04 20:29:25 crc kubenswrapper[4733]: I1204 20:29:25.912470 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ccbe2eef-52a5-41b9-9000-bb0376d74bdc/manila-scheduler/0.log" Dec 04 20:29:25 crc kubenswrapper[4733]: I1204 20:29:25.973183 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d0ee1147-36ba-4d5e-95eb-17e8f8007e23/manila-share/0.log" Dec 04 20:29:26 crc kubenswrapper[4733]: I1204 20:29:26.042686 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d0ee1147-36ba-4d5e-95eb-17e8f8007e23/probe/0.log" Dec 04 20:29:26 crc kubenswrapper[4733]: I1204 20:29:26.368654 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-67fb4dd6bf-g6962_e23aec62-3331-416d-868e-a4ed6b471984/neutron-httpd/0.log" Dec 04 20:29:26 crc kubenswrapper[4733]: I1204 20:29:26.453361 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-67fb4dd6bf-g6962_e23aec62-3331-416d-868e-a4ed6b471984/neutron-api/0.log" Dec 04 20:29:26 crc kubenswrapper[4733]: I1204 20:29:26.487718 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-h9s8f_4bdd1faa-153d-4a64-885a-5f10370c668e/neutron-dhcp-openstack-openstack-cell1/0.log" Dec 04 20:29:26 crc kubenswrapper[4733]: I1204 20:29:26.702127 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-kfjpj_269e8fa2-f463-405c-8a43-8c083f6e6ea9/neutron-metadata-openstack-openstack-cell1/0.log" Dec 04 20:29:26 crc kubenswrapper[4733]: I1204 20:29:26.776324 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-jgtmh_af2a17da-c85e-42d9-ab67-5c44b5c5eab8/neutron-sriov-openstack-openstack-cell1/0.log" Dec 04 20:29:27 crc kubenswrapper[4733]: I1204 20:29:27.006190 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_04fab59a-c08a-4fd3-926a-51a6bd6e6d3f/nova-api-api/0.log" Dec 04 20:29:27 crc kubenswrapper[4733]: I1204 20:29:27.144089 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_04fab59a-c08a-4fd3-926a-51a6bd6e6d3f/nova-api-log/0.log" Dec 04 20:29:27 crc kubenswrapper[4733]: I1204 20:29:27.347724 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_844a3369-74af-4519-919b-f55b202c9e5e/nova-cell0-conductor-conductor/0.log" Dec 04 20:29:27 crc kubenswrapper[4733]: I1204 20:29:27.472533 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ef56127d-af29-4370-9c9d-c2b72b040fac/nova-cell1-conductor-conductor/0.log" Dec 04 20:29:27 crc kubenswrapper[4733]: I1204 20:29:27.581593 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_bfa626a2-12b3-4774-b1f5-404753b13f2e/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 20:29:27 crc kubenswrapper[4733]: I1204 20:29:27.982921 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellx2k5k_5d8da482-a612-4e66-a0f7-7b4f48901cac/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Dec 04 20:29:28 crc kubenswrapper[4733]: I1204 20:29:28.121506 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-lzzhp_97de9284-f0f7-4adf-8f5e-35e3223bd2f8/nova-cell1-openstack-openstack-cell1/0.log" Dec 04 20:29:28 crc kubenswrapper[4733]: I1204 20:29:28.198159 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d50035b9-6d8a-4760-8088-2a9771cb63d1/nova-metadata-metadata/0.log" Dec 04 20:29:28 crc kubenswrapper[4733]: I1204 20:29:28.295594 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d50035b9-6d8a-4760-8088-2a9771cb63d1/nova-metadata-log/0.log" Dec 04 20:29:28 crc kubenswrapper[4733]: I1204 20:29:28.413250 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d1599c17-6cbd-45fe-bfa0-b7862341e8d6/nova-scheduler-scheduler/0.log" Dec 04 20:29:28 crc kubenswrapper[4733]: I1204 20:29:28.486883 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-59457b6897-wl54k_a777e880-541c-4aae-8249-d2027129051f/init/0.log" Dec 04 20:29:28 crc kubenswrapper[4733]: I1204 20:29:28.760521 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-59457b6897-wl54k_a777e880-541c-4aae-8249-d2027129051f/init/0.log" Dec 04 20:29:28 crc kubenswrapper[4733]: I1204 20:29:28.788290 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-59457b6897-wl54k_a777e880-541c-4aae-8249-d2027129051f/octavia-api-provider-agent/0.log" Dec 04 20:29:28 crc kubenswrapper[4733]: I1204 20:29:28.963488 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-mj4sc_5114b53c-38d9-45b5-9f9b-50713a9f8abd/init/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.002326 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-59457b6897-wl54k_a777e880-541c-4aae-8249-d2027129051f/octavia-api/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.200500 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-85ctw_2d9abffe-6c27-4cf4-b943-3ad0f694228f/init/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.245824 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-mj4sc_5114b53c-38d9-45b5-9f9b-50713a9f8abd/octavia-healthmanager/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.283470 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-mj4sc_5114b53c-38d9-45b5-9f9b-50713a9f8abd/init/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.457510 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-85ctw_2d9abffe-6c27-4cf4-b943-3ad0f694228f/init/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.552298 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-xt2bg_8591138e-6ad7-491b-9e0d-45e6e50132de/init/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.582345 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-85ctw_2d9abffe-6c27-4cf4-b943-3ad0f694228f/octavia-housekeeping/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.729911 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-xt2bg_8591138e-6ad7-491b-9e0d-45e6e50132de/init/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.795190 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-x47tj_4ba8c639-2dcf-48f5-a004-0f5e16aee627/init/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.856338 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-xt2bg_8591138e-6ad7-491b-9e0d-45e6e50132de/octavia-rsyslog/0.log" Dec 04 20:29:29 crc kubenswrapper[4733]: I1204 20:29:29.993768 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-x47tj_4ba8c639-2dcf-48f5-a004-0f5e16aee627/init/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.110613 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f7133c01-47a1-4adc-9d8e-2a314124aa03/mysql-bootstrap/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.202707 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-x47tj_4ba8c639-2dcf-48f5-a004-0f5e16aee627/octavia-worker/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.287890 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f7133c01-47a1-4adc-9d8e-2a314124aa03/mysql-bootstrap/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.300319 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f7133c01-47a1-4adc-9d8e-2a314124aa03/galera/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.456679 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5c74c10-5554-486b-b79d-af7a2fa49019/mysql-bootstrap/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.683249 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5c74c10-5554-486b-b79d-af7a2fa49019/mysql-bootstrap/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.695537 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_4dcf8409-c5dd-4cf4-bffd-c08a965fca43/openstackclient/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.702398 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e5c74c10-5554-486b-b79d-af7a2fa49019/galera/0.log" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.780045 4733 scope.go:117] "RemoveContainer" containerID="2010b5b375dbd0132eb5a265d4dcceae05599b30cfe29333b638bfd535719c55" Dec 04 20:29:30 crc kubenswrapper[4733]: I1204 20:29:30.956252 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jnvbm_9ee5eee1-8eae-4de7-a2df-9ebbea833016/ovn-controller/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.046971 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-wwr9r_5187cdcf-f448-48b3-8e84-6e5cd6b7ddc1/openstack-network-exporter/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.168583 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-h6kk4_0d759993-4494-4925-a9cf-97897f8c1383/ovsdb-server-init/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.363932 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-h6kk4_0d759993-4494-4925-a9cf-97897f8c1383/ovs-vswitchd/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.416412 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-h6kk4_0d759993-4494-4925-a9cf-97897f8c1383/ovsdb-server/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.420746 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-h6kk4_0d759993-4494-4925-a9cf-97897f8c1383/ovsdb-server-init/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.586777 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0/openstack-network-exporter/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.626199 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5318c9c7-4e2e-4b2d-9a88-a9611f2c9da0/ovn-northd/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.798766 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_fd6238cc-9a10-42fb-bd5c-f22151dfcd87/openstack-network-exporter/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.850018 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-ljcx8_a11a6634-6fde-4e35-9d88-a7def887dd5b/ovn-openstack-openstack-cell1/0.log" Dec 04 20:29:31 crc kubenswrapper[4733]: I1204 20:29:31.975149 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_fd6238cc-9a10-42fb-bd5c-f22151dfcd87/ovsdbserver-nb/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.168337 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3/openstack-network-exporter/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.175913 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_fb48f9f0-db9c-4f91-801b-e80f1bd2f7c3/ovsdbserver-nb/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.422766 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_fe73daf3-4a24-4813-b103-2b833c958140/openstack-network-exporter/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.479013 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c88aa772-44c8-469c-a29f-eb9d3ae11b9b/openstack-network-exporter/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.503200 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_fe73daf3-4a24-4813-b103-2b833c958140/ovsdbserver-nb/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.593759 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c88aa772-44c8-469c-a29f-eb9d3ae11b9b/ovsdbserver-sb/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.719872 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_18de617e-b3eb-4a20-a530-14476a9de2e4/openstack-network-exporter/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.810904 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_18de617e-b3eb-4a20-a530-14476a9de2e4/ovsdbserver-sb/0.log" Dec 04 20:29:32 crc kubenswrapper[4733]: I1204 20:29:32.987291 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_bd4e8124-0462-4c77-bc59-fb83e7ee2b25/openstack-network-exporter/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.002727 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_bd4e8124-0462-4c77-bc59-fb83e7ee2b25/ovsdbserver-sb/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.104199 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74cddd8d8b-8hsq8_704551cd-6b64-48c0-927c-2b3a857f07a8/placement-api/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.278948 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74cddd8d8b-8hsq8_704551cd-6b64-48c0-927c-2b3a857f07a8/placement-log/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.319234 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-c9krft_70ce3bdd-8a94-4522-a977-106089e82c98/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.476729 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf/init-config-reloader/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.682093 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf/prometheus/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.695981 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf/config-reloader/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.740297 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf/thanos-sidecar/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.745080 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_aede4a3c-0a99-45a2-a1cc-833a9f8cb6bf/init-config-reloader/0.log" Dec 04 20:29:33 crc kubenswrapper[4733]: I1204 20:29:33.937089 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f9f9453e-eacc-4174-b5d5-30ccd949ee2b/setup-container/0.log" Dec 04 20:29:34 crc kubenswrapper[4733]: I1204 20:29:34.185217 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f9f9453e-eacc-4174-b5d5-30ccd949ee2b/setup-container/0.log" Dec 04 20:29:34 crc kubenswrapper[4733]: I1204 20:29:34.189154 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f9f9453e-eacc-4174-b5d5-30ccd949ee2b/rabbitmq/0.log" Dec 04 20:29:34 crc kubenswrapper[4733]: I1204 20:29:34.235050 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9eb02384-3d2d-4977-9896-b57767861dbb/setup-container/0.log" Dec 04 20:29:34 crc kubenswrapper[4733]: I1204 20:29:34.424272 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9eb02384-3d2d-4977-9896-b57767861dbb/setup-container/0.log" Dec 04 20:29:34 crc kubenswrapper[4733]: I1204 20:29:34.595245 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-ck6bp_52516c81-59f5-4b50-9b16-80fcc8303cc9/reboot-os-openstack-openstack-cell1/0.log" Dec 04 20:29:34 crc kubenswrapper[4733]: I1204 20:29:34.841631 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-7ddvl_3ac10d63-265c-49a1-a66c-bd638c729159/run-os-openstack-openstack-cell1/0.log" Dec 04 20:29:34 crc kubenswrapper[4733]: I1204 20:29:34.867114 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9eb02384-3d2d-4977-9896-b57767861dbb/rabbitmq/0.log" Dec 04 20:29:34 crc kubenswrapper[4733]: I1204 20:29:34.937620 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-tnzhn_f485e10f-8830-4d98-ad15-ff39e157e28c/ssh-known-hosts-openstack/0.log" Dec 04 20:29:35 crc kubenswrapper[4733]: I1204 20:29:35.157671 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-hhgzn_469f9bc6-2ba9-4b56-8b58-6a1d5907c1ed/telemetry-openstack-openstack-cell1/0.log" Dec 04 20:29:35 crc kubenswrapper[4733]: I1204 20:29:35.298141 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-shqk2_3dacba4e-76b0-40f9-92b8-67feb437f1a7/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Dec 04 20:29:35 crc kubenswrapper[4733]: I1204 20:29:35.382780 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-s5fzv_32a595da-5163-4f68-99f5-195b933e51f9/validate-network-openstack-openstack-cell1/0.log" Dec 04 20:29:36 crc kubenswrapper[4733]: I1204 20:29:36.639630 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_b365fa5f-abce-4758-a4f2-1307febe0243/memcached/0.log" Dec 04 20:29:57 crc kubenswrapper[4733]: I1204 20:29:57.264747 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx_5e268882-1f3e-40a2-a37c-2ad332106598/util/0.log" Dec 04 20:29:57 crc kubenswrapper[4733]: I1204 20:29:57.590448 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx_5e268882-1f3e-40a2-a37c-2ad332106598/util/0.log" Dec 04 20:29:57 crc kubenswrapper[4733]: I1204 20:29:57.604072 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx_5e268882-1f3e-40a2-a37c-2ad332106598/pull/0.log" Dec 04 20:29:57 crc kubenswrapper[4733]: I1204 20:29:57.605065 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx_5e268882-1f3e-40a2-a37c-2ad332106598/pull/0.log" Dec 04 20:29:57 crc kubenswrapper[4733]: I1204 20:29:57.717437 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx_5e268882-1f3e-40a2-a37c-2ad332106598/util/0.log" Dec 04 20:29:57 crc kubenswrapper[4733]: I1204 20:29:57.776529 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx_5e268882-1f3e-40a2-a37c-2ad332106598/pull/0.log" Dec 04 20:29:57 crc kubenswrapper[4733]: I1204 20:29:57.864550 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafs6qlx_5e268882-1f3e-40a2-a37c-2ad332106598/extract/0.log" Dec 04 20:29:57 crc kubenswrapper[4733]: I1204 20:29:57.972577 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-m45ps_9e110407-2b5b-4f7a-b815-77203c99b56b/kube-rbac-proxy/0.log" Dec 04 20:29:58 crc kubenswrapper[4733]: I1204 20:29:58.094876 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-m45ps_9e110407-2b5b-4f7a-b815-77203c99b56b/manager/0.log" Dec 04 20:29:58 crc kubenswrapper[4733]: I1204 20:29:58.142370 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-8t88q_500a0dd4-904a-449e-956a-013bed8c66ca/kube-rbac-proxy/0.log" Dec 04 20:29:58 crc kubenswrapper[4733]: I1204 20:29:58.248186 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-8t88q_500a0dd4-904a-449e-956a-013bed8c66ca/manager/0.log" Dec 04 20:29:58 crc kubenswrapper[4733]: I1204 20:29:58.382141 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-cs6cz_056b1e59-a544-4274-ae2e-b3d35ea1ab06/kube-rbac-proxy/0.log" Dec 04 20:29:58 crc kubenswrapper[4733]: I1204 20:29:58.948010 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-bjzvv_fbad1f3f-55ee-4477-a8a4-c146ee09ba51/kube-rbac-proxy/0.log" Dec 04 20:29:58 crc kubenswrapper[4733]: I1204 20:29:58.953240 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-cs6cz_056b1e59-a544-4274-ae2e-b3d35ea1ab06/manager/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.094462 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-bjzvv_fbad1f3f-55ee-4477-a8a4-c146ee09ba51/manager/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.193522 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-25jzc_f4fddbf6-0bc5-4b36-9a8a-7073e875c540/kube-rbac-proxy/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.262914 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-25jzc_f4fddbf6-0bc5-4b36-9a8a-7073e875c540/manager/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.386080 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2pm4z_2191192e-388c-4535-be27-5b4b76aea175/kube-rbac-proxy/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.390759 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2pm4z_2191192e-388c-4535-be27-5b4b76aea175/manager/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.549613 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-8skx7_504dcc38-54b7-430f-8c32-e11e5bbbf5dc/kube-rbac-proxy/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.747949 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-v8d4h_da343538-3e4a-49a0-949d-2622d62ed6a2/kube-rbac-proxy/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.829131 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-v8d4h_da343538-3e4a-49a0-949d-2622d62ed6a2/manager/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.901321 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-8skx7_504dcc38-54b7-430f-8c32-e11e5bbbf5dc/manager/0.log" Dec 04 20:29:59 crc kubenswrapper[4733]: I1204 20:29:59.973433 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qpssp_50dfe63c-b3b2-46c5-ad8d-91c64e3ef698/kube-rbac-proxy/0.log" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.111373 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-8r2s2_7f29393d-5068-4870-86ab-237c52f6d1f9/kube-rbac-proxy/0.log" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.143028 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qpssp_50dfe63c-b3b2-46c5-ad8d-91c64e3ef698/manager/0.log" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.162534 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s"] Dec 04 20:30:00 crc kubenswrapper[4733]: E1204 20:30:00.163278 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerName="registry-server" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.163365 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerName="registry-server" Dec 04 20:30:00 crc kubenswrapper[4733]: E1204 20:30:00.163433 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerName="extract-content" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.163491 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerName="extract-content" Dec 04 20:30:00 crc kubenswrapper[4733]: E1204 20:30:00.163580 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerName="extract-utilities" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.163636 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerName="extract-utilities" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.163902 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b604bde3-fe9e-43a9-9c02-fd645889b821" containerName="registry-server" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.165435 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.167937 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.172489 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.190809 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s"] Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.263995 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-8r2s2_7f29393d-5068-4870-86ab-237c52f6d1f9/manager/0.log" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.359864 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-secret-volume\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.359935 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwbnm\" (UniqueName: \"kubernetes.io/projected/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-kube-api-access-hwbnm\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.361001 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-config-volume\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.464542 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-config-volume\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.464708 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-secret-volume\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.464744 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwbnm\" (UniqueName: \"kubernetes.io/projected/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-kube-api-access-hwbnm\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.466724 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-config-volume\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.742258 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-secret-volume\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.743742 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwbnm\" (UniqueName: \"kubernetes.io/projected/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-kube-api-access-hwbnm\") pod \"collect-profiles-29414670-wrd9s\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:00 crc kubenswrapper[4733]: I1204 20:30:00.787946 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.057433 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vqnt5_143e7c55-1ecb-4efa-8992-332718ae5b61/kube-rbac-proxy/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.176194 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vqnt5_143e7c55-1ecb-4efa-8992-332718ae5b61/manager/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.210066 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-txjrm_e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5/kube-rbac-proxy/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.305622 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5rln8_dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8/kube-rbac-proxy/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.312670 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-txjrm_e1164d71-5fbe-4abc-9ec1-f3f41ec0ddd5/manager/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.339338 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s"] Dec 04 20:30:01 crc kubenswrapper[4733]: W1204 20:30:01.348356 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6e647a_b08a_4609_afb6_a2f8cd309bf0.slice/crio-27ded08a754ba7a2724b9656eb82a15de24a4ba3068b8aa996668e2643679f33 WatchSource:0}: Error finding container 27ded08a754ba7a2724b9656eb82a15de24a4ba3068b8aa996668e2643679f33: Status 404 returned error can't find the container with id 27ded08a754ba7a2724b9656eb82a15de24a4ba3068b8aa996668e2643679f33 Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.584695 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-5rln8_dddede3b-d7b1-4ea8-b94a-4646bdbfb1b8/manager/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.594337 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-k5gsb_889320d7-9fc9-43b9-be8b-d5d8ceeb966c/kube-rbac-proxy/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.637692 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-k5gsb_889320d7-9fc9-43b9-be8b-d5d8ceeb966c/manager/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.656464 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" event={"ID":"7a6e647a-b08a-4609-afb6-a2f8cd309bf0","Type":"ContainerStarted","Data":"27ded08a754ba7a2724b9656eb82a15de24a4ba3068b8aa996668e2643679f33"} Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.718511 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f59cdsz_38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee/kube-rbac-proxy/0.log" Dec 04 20:30:01 crc kubenswrapper[4733]: I1204 20:30:01.850945 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f59cdsz_38cad43f-c4ed-4dcf-8de2-6f86aa9ba4ee/manager/0.log" Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.267270 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55b6fb9447-47stg_93a96c03-299d-47fa-9873-2c890b731890/operator/0.log" Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.344817 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fmd92_0f74bf5a-9575-45d2-a08e-ac35a429ecdb/registry-server/0.log" Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.379688 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-v8s48_d0217980-ef7c-4b3a-b5cf-9887545b7a13/kube-rbac-proxy/0.log" Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.541677 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7qmwh_6c895222-dbc2-4490-972c-b7e9dad9e839/kube-rbac-proxy/0.log" Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.588023 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-v8s48_d0217980-ef7c-4b3a-b5cf-9887545b7a13/manager/0.log" Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.671807 4733 generic.go:334] "Generic (PLEG): container finished" podID="7a6e647a-b08a-4609-afb6-a2f8cd309bf0" containerID="96162057b157a3e5369671dc4c3f4cb775e63b03445aad7fe8990a2b8c701728" exitCode=0 Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.671853 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" event={"ID":"7a6e647a-b08a-4609-afb6-a2f8cd309bf0","Type":"ContainerDied","Data":"96162057b157a3e5369671dc4c3f4cb775e63b03445aad7fe8990a2b8c701728"} Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.702782 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7qmwh_6c895222-dbc2-4490-972c-b7e9dad9e839/manager/0.log" Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.822319 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qnndl_65277cff-497a-4135-988c-94d7456a5dfc/operator/0.log" Dec 04 20:30:02 crc kubenswrapper[4733]: I1204 20:30:02.896765 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-7rb9w_89feaf19-9df3-415a-8bbf-57168c3b89ba/kube-rbac-proxy/0.log" Dec 04 20:30:03 crc kubenswrapper[4733]: I1204 20:30:03.110890 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-zw57g_0b437634-e4b9-4e70-b501-2aee9670efea/kube-rbac-proxy/0.log" Dec 04 20:30:03 crc kubenswrapper[4733]: I1204 20:30:03.135778 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-7rb9w_89feaf19-9df3-415a-8bbf-57168c3b89ba/manager/0.log" Dec 04 20:30:03 crc kubenswrapper[4733]: I1204 20:30:03.335657 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z6qpg_eec4a71f-5a1e-493f-b820-523ac0712a46/kube-rbac-proxy/0.log" Dec 04 20:30:03 crc kubenswrapper[4733]: I1204 20:30:03.390782 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-zw57g_0b437634-e4b9-4e70-b501-2aee9670efea/manager/0.log" Dec 04 20:30:03 crc kubenswrapper[4733]: I1204 20:30:03.410562 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z6qpg_eec4a71f-5a1e-493f-b820-523ac0712a46/manager/0.log" Dec 04 20:30:03 crc kubenswrapper[4733]: I1204 20:30:03.628471 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-mf6w7_b68334f1-9e8d-48c1-90d0-f650a6c952ac/kube-rbac-proxy/0.log" Dec 04 20:30:03 crc kubenswrapper[4733]: I1204 20:30:03.660151 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-mf6w7_b68334f1-9e8d-48c1-90d0-f650a6c952ac/manager/0.log" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.164714 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.342162 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwbnm\" (UniqueName: \"kubernetes.io/projected/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-kube-api-access-hwbnm\") pod \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.342281 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-secret-volume\") pod \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.342423 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-config-volume\") pod \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\" (UID: \"7a6e647a-b08a-4609-afb6-a2f8cd309bf0\") " Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.343622 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-config-volume" (OuterVolumeSpecName: "config-volume") pod "7a6e647a-b08a-4609-afb6-a2f8cd309bf0" (UID: "7a6e647a-b08a-4609-afb6-a2f8cd309bf0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.349410 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7a6e647a-b08a-4609-afb6-a2f8cd309bf0" (UID: "7a6e647a-b08a-4609-afb6-a2f8cd309bf0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.351395 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-kube-api-access-hwbnm" (OuterVolumeSpecName: "kube-api-access-hwbnm") pod "7a6e647a-b08a-4609-afb6-a2f8cd309bf0" (UID: "7a6e647a-b08a-4609-afb6-a2f8cd309bf0"). InnerVolumeSpecName "kube-api-access-hwbnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.412256 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-54bdf956c4-gblbb_f2c5de87-7240-4169-8acb-d34261d2479f/manager/0.log" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.445250 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.445286 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwbnm\" (UniqueName: \"kubernetes.io/projected/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-kube-api-access-hwbnm\") on node \"crc\" DevicePath \"\"" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.445299 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a6e647a-b08a-4609-afb6-a2f8cd309bf0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.691524 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" event={"ID":"7a6e647a-b08a-4609-afb6-a2f8cd309bf0","Type":"ContainerDied","Data":"27ded08a754ba7a2724b9656eb82a15de24a4ba3068b8aa996668e2643679f33"} Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.691566 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27ded08a754ba7a2724b9656eb82a15de24a4ba3068b8aa996668e2643679f33" Dec 04 20:30:04 crc kubenswrapper[4733]: I1204 20:30:04.691611 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414670-wrd9s" Dec 04 20:30:05 crc kubenswrapper[4733]: I1204 20:30:05.249864 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn"] Dec 04 20:30:05 crc kubenswrapper[4733]: I1204 20:30:05.257839 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414625-wg9xn"] Dec 04 20:30:06 crc kubenswrapper[4733]: I1204 20:30:06.349207 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14e5bf5e-eafc-4c9a-9a41-e1608463cf0e" path="/var/lib/kubelet/pods/14e5bf5e-eafc-4c9a-9a41-e1608463cf0e/volumes" Dec 04 20:30:25 crc kubenswrapper[4733]: I1204 20:30:25.387489 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-v7zc2_7767c7f9-42fd-4525-903b-3e6073dd694f/control-plane-machine-set-operator/0.log" Dec 04 20:30:25 crc kubenswrapper[4733]: I1204 20:30:25.584635 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-w9vhc_c26584cc-3f5a-47ca-91ae-4f457eb7947e/kube-rbac-proxy/0.log" Dec 04 20:30:25 crc kubenswrapper[4733]: I1204 20:30:25.618968 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-w9vhc_c26584cc-3f5a-47ca-91ae-4f457eb7947e/machine-api-operator/0.log" Dec 04 20:30:30 crc kubenswrapper[4733]: I1204 20:30:30.860859 4733 scope.go:117] "RemoveContainer" containerID="a1d0df7b29d4491e5bbacd2591672d81084bfd407579d9aa4fb00e8515a7d427" Dec 04 20:30:39 crc kubenswrapper[4733]: I1204 20:30:39.605647 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-2cpvp_1f3587c5-c7c0-40e2-855a-3b42ed4088c9/cert-manager-controller/0.log" Dec 04 20:30:39 crc kubenswrapper[4733]: I1204 20:30:39.785600 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-wcdpm_f2b4ae75-af8e-4f85-9139-2f4841ca4f94/cert-manager-cainjector/0.log" Dec 04 20:30:39 crc kubenswrapper[4733]: I1204 20:30:39.845532 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-k5b27_8f22d997-ee5a-462e-8452-d493e4584761/cert-manager-webhook/0.log" Dec 04 20:30:53 crc kubenswrapper[4733]: I1204 20:30:53.455646 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-z84kw_ce10fe7f-07b2-4f30-a52c-1fc47a486ff3/nmstate-console-plugin/0.log" Dec 04 20:30:53 crc kubenswrapper[4733]: I1204 20:30:53.583779 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-l2pgf_cbba5431-8bfc-4fe8-b496-29993e49d748/nmstate-handler/0.log" Dec 04 20:30:53 crc kubenswrapper[4733]: I1204 20:30:53.587978 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-grw9m_3c0e12d4-71fa-444b-a3e7-b47c798053f5/kube-rbac-proxy/0.log" Dec 04 20:30:53 crc kubenswrapper[4733]: I1204 20:30:53.618210 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-grw9m_3c0e12d4-71fa-444b-a3e7-b47c798053f5/nmstate-metrics/0.log" Dec 04 20:30:53 crc kubenswrapper[4733]: I1204 20:30:53.766091 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-7rgd7_57ea0cda-c1a7-400c-89b5-350c5f6a1bb7/nmstate-operator/0.log" Dec 04 20:30:53 crc kubenswrapper[4733]: I1204 20:30:53.817475 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-v2whn_a5af3730-230a-4686-8262-bef9e42c60b5/nmstate-webhook/0.log" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.112605 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pkk28"] Dec 04 20:30:58 crc kubenswrapper[4733]: E1204 20:30:58.113670 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6e647a-b08a-4609-afb6-a2f8cd309bf0" containerName="collect-profiles" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.113684 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6e647a-b08a-4609-afb6-a2f8cd309bf0" containerName="collect-profiles" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.113949 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6e647a-b08a-4609-afb6-a2f8cd309bf0" containerName="collect-profiles" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.115640 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.128245 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pkk28"] Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.237933 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzdn2\" (UniqueName: \"kubernetes.io/projected/ad1057d3-900e-458d-a9f8-115be82f3038-kube-api-access-bzdn2\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.238042 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-utilities\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.238086 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-catalog-content\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.340351 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzdn2\" (UniqueName: \"kubernetes.io/projected/ad1057d3-900e-458d-a9f8-115be82f3038-kube-api-access-bzdn2\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.340729 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-utilities\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.340884 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-catalog-content\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.341824 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-catalog-content\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.341898 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-utilities\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.363584 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzdn2\" (UniqueName: \"kubernetes.io/projected/ad1057d3-900e-458d-a9f8-115be82f3038-kube-api-access-bzdn2\") pod \"redhat-marketplace-pkk28\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.437583 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:30:58 crc kubenswrapper[4733]: I1204 20:30:58.951271 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pkk28"] Dec 04 20:30:59 crc kubenswrapper[4733]: I1204 20:30:59.319764 4733 generic.go:334] "Generic (PLEG): container finished" podID="ad1057d3-900e-458d-a9f8-115be82f3038" containerID="65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e" exitCode=0 Dec 04 20:30:59 crc kubenswrapper[4733]: I1204 20:30:59.319819 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pkk28" event={"ID":"ad1057d3-900e-458d-a9f8-115be82f3038","Type":"ContainerDied","Data":"65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e"} Dec 04 20:30:59 crc kubenswrapper[4733]: I1204 20:30:59.320083 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pkk28" event={"ID":"ad1057d3-900e-458d-a9f8-115be82f3038","Type":"ContainerStarted","Data":"22f96efdf541559346456fc87e338bf0af5637cf1f5dc4b06af0a984d7ca0413"} Dec 04 20:30:59 crc kubenswrapper[4733]: I1204 20:30:59.321542 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 20:31:01 crc kubenswrapper[4733]: I1204 20:31:01.343605 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pkk28" event={"ID":"ad1057d3-900e-458d-a9f8-115be82f3038","Type":"ContainerStarted","Data":"4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55"} Dec 04 20:31:02 crc kubenswrapper[4733]: I1204 20:31:02.356561 4733 generic.go:334] "Generic (PLEG): container finished" podID="ad1057d3-900e-458d-a9f8-115be82f3038" containerID="4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55" exitCode=0 Dec 04 20:31:02 crc kubenswrapper[4733]: I1204 20:31:02.356643 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pkk28" event={"ID":"ad1057d3-900e-458d-a9f8-115be82f3038","Type":"ContainerDied","Data":"4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55"} Dec 04 20:31:03 crc kubenswrapper[4733]: I1204 20:31:03.368774 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pkk28" event={"ID":"ad1057d3-900e-458d-a9f8-115be82f3038","Type":"ContainerStarted","Data":"443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094"} Dec 04 20:31:03 crc kubenswrapper[4733]: I1204 20:31:03.393289 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pkk28" podStartSLOduration=1.917598598 podStartE2EDuration="5.393271158s" podCreationTimestamp="2025-12-04 20:30:58 +0000 UTC" firstStartedPulling="2025-12-04 20:30:59.321338996 +0000 UTC m=+10321.276700042" lastFinishedPulling="2025-12-04 20:31:02.797011566 +0000 UTC m=+10324.752372602" observedRunningTime="2025-12-04 20:31:03.386922256 +0000 UTC m=+10325.342283302" watchObservedRunningTime="2025-12-04 20:31:03.393271158 +0000 UTC m=+10325.348632204" Dec 04 20:31:08 crc kubenswrapper[4733]: I1204 20:31:08.437902 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:31:08 crc kubenswrapper[4733]: I1204 20:31:08.438484 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:31:08 crc kubenswrapper[4733]: I1204 20:31:08.494646 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.325529 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-z54dn_d1b22abb-0421-41f8-a18e-89ddc9685791/kube-rbac-proxy/0.log" Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.483359 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.502528 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-frr-files/0.log" Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.543506 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pkk28"] Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.757125 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-frr-files/0.log" Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.789064 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-z54dn_d1b22abb-0421-41f8-a18e-89ddc9685791/controller/0.log" Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.805163 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-metrics/0.log" Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.817759 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-reloader/0.log" Dec 04 20:31:09 crc kubenswrapper[4733]: I1204 20:31:09.956419 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-reloader/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.083024 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-frr-files/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.149602 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-metrics/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.158053 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-reloader/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.167163 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-metrics/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.326105 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-frr-files/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.361787 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-reloader/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.370716 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/controller/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.387545 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/cp-metrics/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.524258 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/frr-metrics/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.572125 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/kube-rbac-proxy/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.648846 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/kube-rbac-proxy-frr/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.737147 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/reloader/0.log" Dec 04 20:31:10 crc kubenswrapper[4733]: I1204 20:31:10.854254 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-nb7dr_f0a7cd9b-3887-4a5f-b7e0-0d2ab1ab604f/frr-k8s-webhook-server/0.log" Dec 04 20:31:11 crc kubenswrapper[4733]: I1204 20:31:11.026582 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-77769ff6c4-bp6ln_3b8339d9-7351-422c-a9f3-4d9859dcff2f/manager/0.log" Dec 04 20:31:11 crc kubenswrapper[4733]: I1204 20:31:11.189346 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7fc6b79947-ljf4g_e3cd2ad2-42a7-42b1-adb6-0d6080922788/webhook-server/0.log" Dec 04 20:31:11 crc kubenswrapper[4733]: I1204 20:31:11.394659 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-lqqdn_c76f0cb9-9087-4c60-8152-4bdccc3862ef/kube-rbac-proxy/0.log" Dec 04 20:31:11 crc kubenswrapper[4733]: I1204 20:31:11.450109 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pkk28" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" containerName="registry-server" containerID="cri-o://443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094" gracePeriod=2 Dec 04 20:31:11 crc kubenswrapper[4733]: E1204 20:31:11.581963 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad1057d3_900e_458d_a9f8_115be82f3038.slice/crio-443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094.scope\": RecentStats: unable to find data in memory cache]" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.069112 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.139508 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-utilities\") pod \"ad1057d3-900e-458d-a9f8-115be82f3038\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.139672 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzdn2\" (UniqueName: \"kubernetes.io/projected/ad1057d3-900e-458d-a9f8-115be82f3038-kube-api-access-bzdn2\") pod \"ad1057d3-900e-458d-a9f8-115be82f3038\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.139728 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-catalog-content\") pod \"ad1057d3-900e-458d-a9f8-115be82f3038\" (UID: \"ad1057d3-900e-458d-a9f8-115be82f3038\") " Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.140590 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-utilities" (OuterVolumeSpecName: "utilities") pod "ad1057d3-900e-458d-a9f8-115be82f3038" (UID: "ad1057d3-900e-458d-a9f8-115be82f3038"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.147521 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad1057d3-900e-458d-a9f8-115be82f3038-kube-api-access-bzdn2" (OuterVolumeSpecName: "kube-api-access-bzdn2") pod "ad1057d3-900e-458d-a9f8-115be82f3038" (UID: "ad1057d3-900e-458d-a9f8-115be82f3038"). InnerVolumeSpecName "kube-api-access-bzdn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.172327 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad1057d3-900e-458d-a9f8-115be82f3038" (UID: "ad1057d3-900e-458d-a9f8-115be82f3038"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.244194 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzdn2\" (UniqueName: \"kubernetes.io/projected/ad1057d3-900e-458d-a9f8-115be82f3038-kube-api-access-bzdn2\") on node \"crc\" DevicePath \"\"" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.244223 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.244233 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad1057d3-900e-458d-a9f8-115be82f3038-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.320103 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-lqqdn_c76f0cb9-9087-4c60-8152-4bdccc3862ef/speaker/0.log" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.463266 4733 generic.go:334] "Generic (PLEG): container finished" podID="ad1057d3-900e-458d-a9f8-115be82f3038" containerID="443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094" exitCode=0 Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.463586 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pkk28" event={"ID":"ad1057d3-900e-458d-a9f8-115be82f3038","Type":"ContainerDied","Data":"443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094"} Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.463614 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pkk28" event={"ID":"ad1057d3-900e-458d-a9f8-115be82f3038","Type":"ContainerDied","Data":"22f96efdf541559346456fc87e338bf0af5637cf1f5dc4b06af0a984d7ca0413"} Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.463631 4733 scope.go:117] "RemoveContainer" containerID="443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.463816 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pkk28" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.506059 4733 scope.go:117] "RemoveContainer" containerID="4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.509928 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pkk28"] Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.523860 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pkk28"] Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.538548 4733 scope.go:117] "RemoveContainer" containerID="65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.586675 4733 scope.go:117] "RemoveContainer" containerID="443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094" Dec 04 20:31:12 crc kubenswrapper[4733]: E1204 20:31:12.587221 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094\": container with ID starting with 443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094 not found: ID does not exist" containerID="443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.587288 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094"} err="failed to get container status \"443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094\": rpc error: code = NotFound desc = could not find container \"443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094\": container with ID starting with 443030fbed0f3a40c1504856b4e4a6cabecc1edacd771d74eaf851b25518f094 not found: ID does not exist" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.587319 4733 scope.go:117] "RemoveContainer" containerID="4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55" Dec 04 20:31:12 crc kubenswrapper[4733]: E1204 20:31:12.587618 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55\": container with ID starting with 4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55 not found: ID does not exist" containerID="4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.587641 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55"} err="failed to get container status \"4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55\": rpc error: code = NotFound desc = could not find container \"4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55\": container with ID starting with 4984368d9d70255e203f7ef97fe4258d5a4f0f1c676c5e753404a442bde0fa55 not found: ID does not exist" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.587661 4733 scope.go:117] "RemoveContainer" containerID="65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e" Dec 04 20:31:12 crc kubenswrapper[4733]: E1204 20:31:12.587878 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e\": container with ID starting with 65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e not found: ID does not exist" containerID="65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e" Dec 04 20:31:12 crc kubenswrapper[4733]: I1204 20:31:12.587897 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e"} err="failed to get container status \"65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e\": rpc error: code = NotFound desc = could not find container \"65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e\": container with ID starting with 65efeab4a8aba4f789b956962cca8871562de45d55eca9894beacaad07e9020e not found: ID does not exist" Dec 04 20:31:13 crc kubenswrapper[4733]: I1204 20:31:13.879259 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mrbzw_1ea286a4-29bd-4270-ba70-f22bed08fe14/frr/0.log" Dec 04 20:31:14 crc kubenswrapper[4733]: I1204 20:31:14.348126 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" path="/var/lib/kubelet/pods/ad1057d3-900e-458d-a9f8-115be82f3038/volumes" Dec 04 20:31:26 crc kubenswrapper[4733]: I1204 20:31:26.925879 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj_4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512/util/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.101031 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj_4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512/util/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.110447 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj_4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512/pull/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.189181 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj_4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512/pull/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.307218 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj_4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512/extract/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.350598 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj_4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512/pull/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.350608 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931afpflj_4aaba4f3-8c8c-41e8-b7ca-1fc389f1f512/util/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.508558 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6_4209b6af-a01b-4f51-8cf7-d5bc796c4a3b/util/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.725448 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6_4209b6af-a01b-4f51-8cf7-d5bc796c4a3b/util/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.786331 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6_4209b6af-a01b-4f51-8cf7-d5bc796c4a3b/pull/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.786395 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6_4209b6af-a01b-4f51-8cf7-d5bc796c4a3b/pull/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.854010 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6_4209b6af-a01b-4f51-8cf7-d5bc796c4a3b/util/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.870318 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6_4209b6af-a01b-4f51-8cf7-d5bc796c4a3b/pull/0.log" Dec 04 20:31:27 crc kubenswrapper[4733]: I1204 20:31:27.970679 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkr6r6_4209b6af-a01b-4f51-8cf7-d5bc796c4a3b/extract/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.060766 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s_d4853046-f3d5-48dc-8bb8-aa019bf731dd/util/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.199143 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s_d4853046-f3d5-48dc-8bb8-aa019bf731dd/pull/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.212736 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s_d4853046-f3d5-48dc-8bb8-aa019bf731dd/pull/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.231697 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s_d4853046-f3d5-48dc-8bb8-aa019bf731dd/util/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.423765 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s_d4853046-f3d5-48dc-8bb8-aa019bf731dd/extract/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.424492 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s_d4853046-f3d5-48dc-8bb8-aa019bf731dd/util/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.440078 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210mwd6s_d4853046-f3d5-48dc-8bb8-aa019bf731dd/pull/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.578398 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct_d54cd1df-e504-4d88-8d0d-41610c92c294/util/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.782693 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct_d54cd1df-e504-4d88-8d0d-41610c92c294/util/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.801203 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct_d54cd1df-e504-4d88-8d0d-41610c92c294/pull/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.815360 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct_d54cd1df-e504-4d88-8d0d-41610c92c294/pull/0.log" Dec 04 20:31:28 crc kubenswrapper[4733]: I1204 20:31:28.998552 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct_d54cd1df-e504-4d88-8d0d-41610c92c294/util/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.022775 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct_d54cd1df-e504-4d88-8d0d-41610c92c294/extract/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.024615 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f837jxct_d54cd1df-e504-4d88-8d0d-41610c92c294/pull/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.173465 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cdbgr_761f493e-8e27-4b0d-9a01-09b53b38e142/extract-utilities/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.337433 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cdbgr_761f493e-8e27-4b0d-9a01-09b53b38e142/extract-content/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.360831 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cdbgr_761f493e-8e27-4b0d-9a01-09b53b38e142/extract-content/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.362106 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cdbgr_761f493e-8e27-4b0d-9a01-09b53b38e142/extract-utilities/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.530502 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cdbgr_761f493e-8e27-4b0d-9a01-09b53b38e142/extract-utilities/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.534465 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cdbgr_761f493e-8e27-4b0d-9a01-09b53b38e142/extract-content/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.853568 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bqhkt_af3a70e7-c70d-46bb-a133-1693b8da280a/extract-utilities/0.log" Dec 04 20:31:29 crc kubenswrapper[4733]: I1204 20:31:29.963118 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bqhkt_af3a70e7-c70d-46bb-a133-1693b8da280a/extract-utilities/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.030420 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bqhkt_af3a70e7-c70d-46bb-a133-1693b8da280a/extract-content/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.069221 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bqhkt_af3a70e7-c70d-46bb-a133-1693b8da280a/extract-content/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.241884 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bqhkt_af3a70e7-c70d-46bb-a133-1693b8da280a/extract-utilities/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.267164 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bqhkt_af3a70e7-c70d-46bb-a133-1693b8da280a/extract-content/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.475006 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-545tk_17e52f55-0856-43c8-912b-6594e33aaebc/marketplace-operator/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.561501 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gb2qr_d1d40421-3817-47d8-85c1-aeb819796b70/extract-utilities/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.769589 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cdbgr_761f493e-8e27-4b0d-9a01-09b53b38e142/registry-server/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.882751 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gb2qr_d1d40421-3817-47d8-85c1-aeb819796b70/extract-content/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.900203 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gb2qr_d1d40421-3817-47d8-85c1-aeb819796b70/extract-utilities/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.912284 4733 scope.go:117] "RemoveContainer" containerID="8b515b5206cbc162275d4b903384638e803f75a652e4f73387fa6369c3fd10d4" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.935740 4733 scope.go:117] "RemoveContainer" containerID="e58fdf3e112c8712c75cfa37903fefa157ca1c56aa9865be50e30424fa5713b2" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.942505 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gb2qr_d1d40421-3817-47d8-85c1-aeb819796b70/extract-content/0.log" Dec 04 20:31:30 crc kubenswrapper[4733]: I1204 20:31:30.991267 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gb2qr_d1d40421-3817-47d8-85c1-aeb819796b70/extract-utilities/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.017497 4733 scope.go:117] "RemoveContainer" containerID="06582b122a58e006b681c03d765940d8c7bfe210941badf944c4310b2340dc22" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.145135 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bqhkt_af3a70e7-c70d-46bb-a133-1693b8da280a/registry-server/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.247556 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gb2qr_d1d40421-3817-47d8-85c1-aeb819796b70/extract-content/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.261767 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tvh6f_b048fb6d-0dcc-4daa-9c9d-8548c6464ee6/extract-utilities/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.557592 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tvh6f_b048fb6d-0dcc-4daa-9c9d-8548c6464ee6/extract-content/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.562329 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tvh6f_b048fb6d-0dcc-4daa-9c9d-8548c6464ee6/extract-utilities/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.578231 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gb2qr_d1d40421-3817-47d8-85c1-aeb819796b70/registry-server/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.606101 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tvh6f_b048fb6d-0dcc-4daa-9c9d-8548c6464ee6/extract-content/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.725661 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tvh6f_b048fb6d-0dcc-4daa-9c9d-8548c6464ee6/extract-utilities/0.log" Dec 04 20:31:31 crc kubenswrapper[4733]: I1204 20:31:31.775664 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tvh6f_b048fb6d-0dcc-4daa-9c9d-8548c6464ee6/extract-content/0.log" Dec 04 20:31:32 crc kubenswrapper[4733]: I1204 20:31:32.054643 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tvh6f_b048fb6d-0dcc-4daa-9c9d-8548c6464ee6/registry-server/0.log" Dec 04 20:31:44 crc kubenswrapper[4733]: I1204 20:31:44.894422 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-5lv7r_1aa893e4-e63a-4f24-97bf-a0c41cf27c64/prometheus-operator/0.log" Dec 04 20:31:45 crc kubenswrapper[4733]: I1204 20:31:45.104397 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-56c5df98f7-5qv7q_37deee19-62e0-46de-bf27-99748ce7a98b/prometheus-operator-admission-webhook/0.log" Dec 04 20:31:45 crc kubenswrapper[4733]: I1204 20:31:45.193744 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-56c5df98f7-mr9fp_7f54ccac-c4f7-47fe-b071-87d8c3231848/prometheus-operator-admission-webhook/0.log" Dec 04 20:31:45 crc kubenswrapper[4733]: I1204 20:31:45.309633 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-pjs6d_97bb703c-e296-4b6d-b878-659c7149d970/operator/0.log" Dec 04 20:31:45 crc kubenswrapper[4733]: I1204 20:31:45.362041 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:31:45 crc kubenswrapper[4733]: I1204 20:31:45.362121 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:31:45 crc kubenswrapper[4733]: I1204 20:31:45.431697 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-95nbh_3e69f6b5-cf40-4ebe-9949-71031eb10fc8/perses-operator/0.log" Dec 04 20:32:15 crc kubenswrapper[4733]: I1204 20:32:15.361960 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:32:15 crc kubenswrapper[4733]: I1204 20:32:15.362628 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:32:45 crc kubenswrapper[4733]: I1204 20:32:45.361898 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:32:45 crc kubenswrapper[4733]: I1204 20:32:45.362726 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:32:45 crc kubenswrapper[4733]: I1204 20:32:45.362792 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 20:32:45 crc kubenswrapper[4733]: I1204 20:32:45.363828 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb18a7d2636425cdb65bf7d0042b43bec6a6390071894b7c5044793db67a3f33"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 20:32:45 crc kubenswrapper[4733]: I1204 20:32:45.363926 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://fb18a7d2636425cdb65bf7d0042b43bec6a6390071894b7c5044793db67a3f33" gracePeriod=600 Dec 04 20:32:46 crc kubenswrapper[4733]: I1204 20:32:46.483299 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="fb18a7d2636425cdb65bf7d0042b43bec6a6390071894b7c5044793db67a3f33" exitCode=0 Dec 04 20:32:46 crc kubenswrapper[4733]: I1204 20:32:46.483660 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"fb18a7d2636425cdb65bf7d0042b43bec6a6390071894b7c5044793db67a3f33"} Dec 04 20:32:46 crc kubenswrapper[4733]: I1204 20:32:46.483702 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerStarted","Data":"e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f"} Dec 04 20:32:46 crc kubenswrapper[4733]: I1204 20:32:46.483729 4733 scope.go:117] "RemoveContainer" containerID="c2fc28a2f1bbd12622fa2f4ff638735c4877068ce3d623c84b36de96d11cdfce" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.039708 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dvp4k"] Dec 04 20:33:19 crc kubenswrapper[4733]: E1204 20:33:19.041852 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" containerName="extract-content" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.041872 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" containerName="extract-content" Dec 04 20:33:19 crc kubenswrapper[4733]: E1204 20:33:19.041952 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" containerName="extract-utilities" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.041961 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" containerName="extract-utilities" Dec 04 20:33:19 crc kubenswrapper[4733]: E1204 20:33:19.042066 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" containerName="registry-server" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.042077 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" containerName="registry-server" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.042428 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad1057d3-900e-458d-a9f8-115be82f3038" containerName="registry-server" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.046349 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.058646 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvp4k"] Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.108385 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b1f0207-acf0-4996-9946-e213fc83bf70-utilities\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.108444 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b1f0207-acf0-4996-9946-e213fc83bf70-catalog-content\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.108571 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw9xd\" (UniqueName: \"kubernetes.io/projected/9b1f0207-acf0-4996-9946-e213fc83bf70-kube-api-access-sw9xd\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.210800 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b1f0207-acf0-4996-9946-e213fc83bf70-utilities\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.211181 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b1f0207-acf0-4996-9946-e213fc83bf70-catalog-content\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.211398 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw9xd\" (UniqueName: \"kubernetes.io/projected/9b1f0207-acf0-4996-9946-e213fc83bf70-kube-api-access-sw9xd\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.211394 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b1f0207-acf0-4996-9946-e213fc83bf70-utilities\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.211632 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b1f0207-acf0-4996-9946-e213fc83bf70-catalog-content\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.230532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw9xd\" (UniqueName: \"kubernetes.io/projected/9b1f0207-acf0-4996-9946-e213fc83bf70-kube-api-access-sw9xd\") pod \"redhat-operators-dvp4k\" (UID: \"9b1f0207-acf0-4996-9946-e213fc83bf70\") " pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.368875 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:19 crc kubenswrapper[4733]: I1204 20:33:19.887378 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvp4k"] Dec 04 20:33:20 crc kubenswrapper[4733]: I1204 20:33:20.878442 4733 generic.go:334] "Generic (PLEG): container finished" podID="9b1f0207-acf0-4996-9946-e213fc83bf70" containerID="23740f0e1f4a7606145926c2c1aaa09dadd888307b1b7ec848d6d70b03dae88d" exitCode=0 Dec 04 20:33:20 crc kubenswrapper[4733]: I1204 20:33:20.878626 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvp4k" event={"ID":"9b1f0207-acf0-4996-9946-e213fc83bf70","Type":"ContainerDied","Data":"23740f0e1f4a7606145926c2c1aaa09dadd888307b1b7ec848d6d70b03dae88d"} Dec 04 20:33:20 crc kubenswrapper[4733]: I1204 20:33:20.879250 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvp4k" event={"ID":"9b1f0207-acf0-4996-9946-e213fc83bf70","Type":"ContainerStarted","Data":"d0be1f6061ca6886cae4cebf4d8a1d272a0fb7f7e9dc0bc99e3ac49fefb42041"} Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.525021 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vmfgs"] Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.528234 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.539199 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vmfgs"] Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.585970 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-utilities\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.586386 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64z7t\" (UniqueName: \"kubernetes.io/projected/0427a02d-fbf6-4674-808a-3314681569bd-kube-api-access-64z7t\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.586482 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-catalog-content\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.688473 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-catalog-content\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.688600 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-utilities\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.688721 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64z7t\" (UniqueName: \"kubernetes.io/projected/0427a02d-fbf6-4674-808a-3314681569bd-kube-api-access-64z7t\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.689083 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-catalog-content\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.689095 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-utilities\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.711144 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64z7t\" (UniqueName: \"kubernetes.io/projected/0427a02d-fbf6-4674-808a-3314681569bd-kube-api-access-64z7t\") pod \"community-operators-vmfgs\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:30 crc kubenswrapper[4733]: I1204 20:33:30.850248 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:32 crc kubenswrapper[4733]: I1204 20:33:32.323485 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vmfgs"] Dec 04 20:33:33 crc kubenswrapper[4733]: I1204 20:33:33.017692 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvp4k" event={"ID":"9b1f0207-acf0-4996-9946-e213fc83bf70","Type":"ContainerStarted","Data":"d40ae3682f20efa622ab55c5eb0f46e9ddfd337e766ee7f6efdddcd9d4f2c3a1"} Dec 04 20:33:33 crc kubenswrapper[4733]: I1204 20:33:33.031369 4733 generic.go:334] "Generic (PLEG): container finished" podID="0427a02d-fbf6-4674-808a-3314681569bd" containerID="5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84" exitCode=0 Dec 04 20:33:33 crc kubenswrapper[4733]: I1204 20:33:33.031663 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmfgs" event={"ID":"0427a02d-fbf6-4674-808a-3314681569bd","Type":"ContainerDied","Data":"5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84"} Dec 04 20:33:33 crc kubenswrapper[4733]: I1204 20:33:33.031778 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmfgs" event={"ID":"0427a02d-fbf6-4674-808a-3314681569bd","Type":"ContainerStarted","Data":"0709d3cf9fed5c1f0ba75814297353cb3962ab1163029e3989c06cf54fa53965"} Dec 04 20:33:36 crc kubenswrapper[4733]: I1204 20:33:36.074259 4733 generic.go:334] "Generic (PLEG): container finished" podID="9b1f0207-acf0-4996-9946-e213fc83bf70" containerID="d40ae3682f20efa622ab55c5eb0f46e9ddfd337e766ee7f6efdddcd9d4f2c3a1" exitCode=0 Dec 04 20:33:36 crc kubenswrapper[4733]: I1204 20:33:36.074350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvp4k" event={"ID":"9b1f0207-acf0-4996-9946-e213fc83bf70","Type":"ContainerDied","Data":"d40ae3682f20efa622ab55c5eb0f46e9ddfd337e766ee7f6efdddcd9d4f2c3a1"} Dec 04 20:33:37 crc kubenswrapper[4733]: I1204 20:33:37.086088 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmfgs" event={"ID":"0427a02d-fbf6-4674-808a-3314681569bd","Type":"ContainerStarted","Data":"b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1"} Dec 04 20:33:38 crc kubenswrapper[4733]: I1204 20:33:38.103576 4733 generic.go:334] "Generic (PLEG): container finished" podID="0427a02d-fbf6-4674-808a-3314681569bd" containerID="b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1" exitCode=0 Dec 04 20:33:38 crc kubenswrapper[4733]: I1204 20:33:38.103641 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmfgs" event={"ID":"0427a02d-fbf6-4674-808a-3314681569bd","Type":"ContainerDied","Data":"b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1"} Dec 04 20:33:39 crc kubenswrapper[4733]: I1204 20:33:39.132080 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvp4k" event={"ID":"9b1f0207-acf0-4996-9946-e213fc83bf70","Type":"ContainerStarted","Data":"04fe9fe44f9b93e57f257f342e4315b2ebba1876e684d72a6f28670154f60d2a"} Dec 04 20:33:39 crc kubenswrapper[4733]: I1204 20:33:39.153968 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dvp4k" podStartSLOduration=4.06229556 podStartE2EDuration="20.153953094s" podCreationTimestamp="2025-12-04 20:33:19 +0000 UTC" firstStartedPulling="2025-12-04 20:33:20.880580361 +0000 UTC m=+10462.835941407" lastFinishedPulling="2025-12-04 20:33:36.972237895 +0000 UTC m=+10478.927598941" observedRunningTime="2025-12-04 20:33:39.150322166 +0000 UTC m=+10481.105683242" watchObservedRunningTime="2025-12-04 20:33:39.153953094 +0000 UTC m=+10481.109314140" Dec 04 20:33:39 crc kubenswrapper[4733]: I1204 20:33:39.369546 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:39 crc kubenswrapper[4733]: I1204 20:33:39.369617 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:40 crc kubenswrapper[4733]: I1204 20:33:40.146139 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmfgs" event={"ID":"0427a02d-fbf6-4674-808a-3314681569bd","Type":"ContainerStarted","Data":"0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999"} Dec 04 20:33:40 crc kubenswrapper[4733]: I1204 20:33:40.176008 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vmfgs" podStartSLOduration=4.189020834 podStartE2EDuration="10.17598645s" podCreationTimestamp="2025-12-04 20:33:30 +0000 UTC" firstStartedPulling="2025-12-04 20:33:33.035603799 +0000 UTC m=+10474.990964835" lastFinishedPulling="2025-12-04 20:33:39.022569375 +0000 UTC m=+10480.977930451" observedRunningTime="2025-12-04 20:33:40.166103171 +0000 UTC m=+10482.121464217" watchObservedRunningTime="2025-12-04 20:33:40.17598645 +0000 UTC m=+10482.131347506" Dec 04 20:33:40 crc kubenswrapper[4733]: I1204 20:33:40.454883 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dvp4k" podUID="9b1f0207-acf0-4996-9946-e213fc83bf70" containerName="registry-server" probeResult="failure" output=< Dec 04 20:33:40 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 20:33:40 crc kubenswrapper[4733]: > Dec 04 20:33:40 crc kubenswrapper[4733]: I1204 20:33:40.851117 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:40 crc kubenswrapper[4733]: I1204 20:33:40.851252 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:41 crc kubenswrapper[4733]: I1204 20:33:41.899025 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-vmfgs" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="registry-server" probeResult="failure" output=< Dec 04 20:33:41 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 20:33:41 crc kubenswrapper[4733]: > Dec 04 20:33:50 crc kubenswrapper[4733]: I1204 20:33:50.417694 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dvp4k" podUID="9b1f0207-acf0-4996-9946-e213fc83bf70" containerName="registry-server" probeResult="failure" output=< Dec 04 20:33:50 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Dec 04 20:33:50 crc kubenswrapper[4733]: > Dec 04 20:33:50 crc kubenswrapper[4733]: I1204 20:33:50.908263 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:50 crc kubenswrapper[4733]: I1204 20:33:50.965607 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:52 crc kubenswrapper[4733]: I1204 20:33:52.718312 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vmfgs"] Dec 04 20:33:52 crc kubenswrapper[4733]: I1204 20:33:52.718955 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vmfgs" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="registry-server" containerID="cri-o://0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999" gracePeriod=2 Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.280473 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.292828 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmfgs" event={"ID":"0427a02d-fbf6-4674-808a-3314681569bd","Type":"ContainerDied","Data":"0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999"} Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.292908 4733 scope.go:117] "RemoveContainer" containerID="0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.292842 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vmfgs" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.292727 4733 generic.go:334] "Generic (PLEG): container finished" podID="0427a02d-fbf6-4674-808a-3314681569bd" containerID="0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999" exitCode=0 Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.293057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmfgs" event={"ID":"0427a02d-fbf6-4674-808a-3314681569bd","Type":"ContainerDied","Data":"0709d3cf9fed5c1f0ba75814297353cb3962ab1163029e3989c06cf54fa53965"} Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.322053 4733 scope.go:117] "RemoveContainer" containerID="b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.348499 4733 scope.go:117] "RemoveContainer" containerID="5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.453009 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64z7t\" (UniqueName: \"kubernetes.io/projected/0427a02d-fbf6-4674-808a-3314681569bd-kube-api-access-64z7t\") pod \"0427a02d-fbf6-4674-808a-3314681569bd\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.453178 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-catalog-content\") pod \"0427a02d-fbf6-4674-808a-3314681569bd\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.453238 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-utilities\") pod \"0427a02d-fbf6-4674-808a-3314681569bd\" (UID: \"0427a02d-fbf6-4674-808a-3314681569bd\") " Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.454197 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-utilities" (OuterVolumeSpecName: "utilities") pod "0427a02d-fbf6-4674-808a-3314681569bd" (UID: "0427a02d-fbf6-4674-808a-3314681569bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.506397 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0427a02d-fbf6-4674-808a-3314681569bd" (UID: "0427a02d-fbf6-4674-808a-3314681569bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.556720 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.556768 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0427a02d-fbf6-4674-808a-3314681569bd-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.946789 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0427a02d-fbf6-4674-808a-3314681569bd-kube-api-access-64z7t" (OuterVolumeSpecName: "kube-api-access-64z7t") pod "0427a02d-fbf6-4674-808a-3314681569bd" (UID: "0427a02d-fbf6-4674-808a-3314681569bd"). InnerVolumeSpecName "kube-api-access-64z7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:33:53 crc kubenswrapper[4733]: I1204 20:33:53.965394 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64z7t\" (UniqueName: \"kubernetes.io/projected/0427a02d-fbf6-4674-808a-3314681569bd-kube-api-access-64z7t\") on node \"crc\" DevicePath \"\"" Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.048315 4733 scope.go:117] "RemoveContainer" containerID="0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999" Dec 04 20:33:54 crc kubenswrapper[4733]: E1204 20:33:54.048863 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999\": container with ID starting with 0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999 not found: ID does not exist" containerID="0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999" Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.048894 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999"} err="failed to get container status \"0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999\": rpc error: code = NotFound desc = could not find container \"0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999\": container with ID starting with 0fb005eae1b011fe64686f8de5dd853e8db8ce188ebcae99bbce7ca718cf5999 not found: ID does not exist" Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.048920 4733 scope.go:117] "RemoveContainer" containerID="b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1" Dec 04 20:33:54 crc kubenswrapper[4733]: E1204 20:33:54.049466 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1\": container with ID starting with b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1 not found: ID does not exist" containerID="b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1" Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.049495 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1"} err="failed to get container status \"b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1\": rpc error: code = NotFound desc = could not find container \"b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1\": container with ID starting with b6f311264ef24678e0e6bfce796319fb6ba447cd128dbeaa5832aa5e9d10e5e1 not found: ID does not exist" Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.049510 4733 scope.go:117] "RemoveContainer" containerID="5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84" Dec 04 20:33:54 crc kubenswrapper[4733]: E1204 20:33:54.049784 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84\": container with ID starting with 5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84 not found: ID does not exist" containerID="5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84" Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.049823 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84"} err="failed to get container status \"5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84\": rpc error: code = NotFound desc = could not find container \"5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84\": container with ID starting with 5fb648a05d090eef70bfb188f6905711a86d69695f5e548ce16e38c66c239c84 not found: ID does not exist" Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.239719 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vmfgs"] Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.249742 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vmfgs"] Dec 04 20:33:54 crc kubenswrapper[4733]: I1204 20:33:54.347687 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0427a02d-fbf6-4674-808a-3314681569bd" path="/var/lib/kubelet/pods/0427a02d-fbf6-4674-808a-3314681569bd/volumes" Dec 04 20:33:59 crc kubenswrapper[4733]: I1204 20:33:59.426710 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:33:59 crc kubenswrapper[4733]: I1204 20:33:59.477597 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dvp4k" Dec 04 20:34:03 crc kubenswrapper[4733]: I1204 20:34:03.139753 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvp4k"] Dec 04 20:34:03 crc kubenswrapper[4733]: I1204 20:34:03.919052 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tvh6f"] Dec 04 20:34:03 crc kubenswrapper[4733]: I1204 20:34:03.919667 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tvh6f" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="registry-server" containerID="cri-o://c791df1d01ca38a8be60d594f44123326c3688546f6097125c45a4ce83fee33e" gracePeriod=2 Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.426368 4733 generic.go:334] "Generic (PLEG): container finished" podID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerID="c791df1d01ca38a8be60d594f44123326c3688546f6097125c45a4ce83fee33e" exitCode=0 Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.426457 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tvh6f" event={"ID":"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6","Type":"ContainerDied","Data":"c791df1d01ca38a8be60d594f44123326c3688546f6097125c45a4ce83fee33e"} Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.426762 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tvh6f" event={"ID":"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6","Type":"ContainerDied","Data":"d4492a43538796c1c099fbf67730f01d7c212b3918810d4468056beee0dd6abd"} Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.426782 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4492a43538796c1c099fbf67730f01d7c212b3918810d4468056beee0dd6abd" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.430040 4733 generic.go:334] "Generic (PLEG): container finished" podID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerID="5926f52b4eb9210fadba47185d2905a28d1a3ef594789f628b6b7275ad423935" exitCode=0 Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.430125 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n24qc/must-gather-phwjg" event={"ID":"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0","Type":"ContainerDied","Data":"5926f52b4eb9210fadba47185d2905a28d1a3ef594789f628b6b7275ad423935"} Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.431267 4733 scope.go:117] "RemoveContainer" containerID="5926f52b4eb9210fadba47185d2905a28d1a3ef594789f628b6b7275ad423935" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.492354 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.640760 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-catalog-content\") pod \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.640873 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-utilities\") pod \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.640897 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttbhz\" (UniqueName: \"kubernetes.io/projected/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-kube-api-access-ttbhz\") pod \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\" (UID: \"b048fb6d-0dcc-4daa-9c9d-8548c6464ee6\") " Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.641418 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-utilities" (OuterVolumeSpecName: "utilities") pod "b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" (UID: "b048fb6d-0dcc-4daa-9c9d-8548c6464ee6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.641746 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.648196 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-kube-api-access-ttbhz" (OuterVolumeSpecName: "kube-api-access-ttbhz") pod "b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" (UID: "b048fb6d-0dcc-4daa-9c9d-8548c6464ee6"). InnerVolumeSpecName "kube-api-access-ttbhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.743188 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttbhz\" (UniqueName: \"kubernetes.io/projected/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-kube-api-access-ttbhz\") on node \"crc\" DevicePath \"\"" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.746239 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n24qc_must-gather-phwjg_a53bb1e9-f3b4-4f84-81ec-9c12067c72e0/gather/0.log" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.761134 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" (UID: "b048fb6d-0dcc-4daa-9c9d-8548c6464ee6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:34:04 crc kubenswrapper[4733]: I1204 20:34:04.845301 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:34:05 crc kubenswrapper[4733]: I1204 20:34:05.441506 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tvh6f" Dec 04 20:34:05 crc kubenswrapper[4733]: I1204 20:34:05.485565 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tvh6f"] Dec 04 20:34:05 crc kubenswrapper[4733]: I1204 20:34:05.497306 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tvh6f"] Dec 04 20:34:06 crc kubenswrapper[4733]: I1204 20:34:06.351582 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" path="/var/lib/kubelet/pods/b048fb6d-0dcc-4daa-9c9d-8548c6464ee6/volumes" Dec 04 20:34:15 crc kubenswrapper[4733]: I1204 20:34:15.804512 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n24qc/must-gather-phwjg"] Dec 04 20:34:15 crc kubenswrapper[4733]: I1204 20:34:15.805364 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-n24qc/must-gather-phwjg" podUID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerName="copy" containerID="cri-o://0496e1082cbe5bd8b0c9759cadf39d18f3a1d2dd09604bb8a9b969589e3c3486" gracePeriod=2 Dec 04 20:34:15 crc kubenswrapper[4733]: I1204 20:34:15.826672 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n24qc/must-gather-phwjg"] Dec 04 20:34:16 crc kubenswrapper[4733]: I1204 20:34:16.557312 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n24qc_must-gather-phwjg_a53bb1e9-f3b4-4f84-81ec-9c12067c72e0/copy/0.log" Dec 04 20:34:16 crc kubenswrapper[4733]: I1204 20:34:16.558128 4733 generic.go:334] "Generic (PLEG): container finished" podID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerID="0496e1082cbe5bd8b0c9759cadf39d18f3a1d2dd09604bb8a9b969589e3c3486" exitCode=143 Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.061678 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n24qc_must-gather-phwjg_a53bb1e9-f3b4-4f84-81ec-9c12067c72e0/copy/0.log" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.062852 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.169709 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc2wf\" (UniqueName: \"kubernetes.io/projected/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-kube-api-access-hc2wf\") pod \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\" (UID: \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\") " Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.169821 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-must-gather-output\") pod \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\" (UID: \"a53bb1e9-f3b4-4f84-81ec-9c12067c72e0\") " Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.177783 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-kube-api-access-hc2wf" (OuterVolumeSpecName: "kube-api-access-hc2wf") pod "a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" (UID: "a53bb1e9-f3b4-4f84-81ec-9c12067c72e0"). InnerVolumeSpecName "kube-api-access-hc2wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.272449 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc2wf\" (UniqueName: \"kubernetes.io/projected/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-kube-api-access-hc2wf\") on node \"crc\" DevicePath \"\"" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.375194 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" (UID: "a53bb1e9-f3b4-4f84-81ec-9c12067c72e0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.480274 4733 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.568579 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n24qc_must-gather-phwjg_a53bb1e9-f3b4-4f84-81ec-9c12067c72e0/copy/0.log" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.568934 4733 scope.go:117] "RemoveContainer" containerID="0496e1082cbe5bd8b0c9759cadf39d18f3a1d2dd09604bb8a9b969589e3c3486" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.569090 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n24qc/must-gather-phwjg" Dec 04 20:34:17 crc kubenswrapper[4733]: I1204 20:34:17.600151 4733 scope.go:117] "RemoveContainer" containerID="5926f52b4eb9210fadba47185d2905a28d1a3ef594789f628b6b7275ad423935" Dec 04 20:34:18 crc kubenswrapper[4733]: I1204 20:34:18.349971 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" path="/var/lib/kubelet/pods/a53bb1e9-f3b4-4f84-81ec-9c12067c72e0/volumes" Dec 04 20:34:31 crc kubenswrapper[4733]: I1204 20:34:31.208625 4733 scope.go:117] "RemoveContainer" containerID="0742b46534d0e8c24ccc96e3490125aa7d0f21d22bdef3fe2321538755076054" Dec 04 20:34:31 crc kubenswrapper[4733]: I1204 20:34:31.262348 4733 scope.go:117] "RemoveContainer" containerID="c791df1d01ca38a8be60d594f44123326c3688546f6097125c45a4ce83fee33e" Dec 04 20:34:31 crc kubenswrapper[4733]: I1204 20:34:31.321978 4733 scope.go:117] "RemoveContainer" containerID="6ac8d28501c04023040ed0ae5abc820413631c7e9a09c246074a8f96d13880a4" Dec 04 20:34:45 crc kubenswrapper[4733]: I1204 20:34:45.362377 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:34:45 crc kubenswrapper[4733]: I1204 20:34:45.363222 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.098940 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6xr4k"] Dec 04 20:34:56 crc kubenswrapper[4733]: E1204 20:34:56.100694 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="registry-server" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.100709 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="registry-server" Dec 04 20:34:56 crc kubenswrapper[4733]: E1204 20:34:56.100738 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="extract-utilities" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.100745 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="extract-utilities" Dec 04 20:34:56 crc kubenswrapper[4733]: E1204 20:34:56.100756 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="extract-utilities" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.100762 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="extract-utilities" Dec 04 20:34:56 crc kubenswrapper[4733]: E1204 20:34:56.100773 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerName="gather" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.100779 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerName="gather" Dec 04 20:34:56 crc kubenswrapper[4733]: E1204 20:34:56.100790 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="extract-content" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.100810 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="extract-content" Dec 04 20:34:56 crc kubenswrapper[4733]: E1204 20:34:56.100824 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="extract-content" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.100829 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="extract-content" Dec 04 20:34:56 crc kubenswrapper[4733]: E1204 20:34:56.100849 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerName="copy" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.100854 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerName="copy" Dec 04 20:34:56 crc kubenswrapper[4733]: E1204 20:34:56.100863 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="registry-server" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.100869 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="registry-server" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.101049 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0427a02d-fbf6-4674-808a-3314681569bd" containerName="registry-server" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.101064 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerName="gather" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.101075 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b048fb6d-0dcc-4daa-9c9d-8548c6464ee6" containerName="registry-server" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.101094 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53bb1e9-f3b4-4f84-81ec-9c12067c72e0" containerName="copy" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.104251 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.113851 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xr4k"] Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.166514 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8f44\" (UniqueName: \"kubernetes.io/projected/0a5b223e-fe0f-4593-9a83-e81b54be6a63-kube-api-access-n8f44\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.166587 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-catalog-content\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.166787 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-utilities\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.268856 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8f44\" (UniqueName: \"kubernetes.io/projected/0a5b223e-fe0f-4593-9a83-e81b54be6a63-kube-api-access-n8f44\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.268949 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-catalog-content\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.269041 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-utilities\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.269683 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-utilities\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.270335 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-catalog-content\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.289836 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8f44\" (UniqueName: \"kubernetes.io/projected/0a5b223e-fe0f-4593-9a83-e81b54be6a63-kube-api-access-n8f44\") pod \"certified-operators-6xr4k\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.427597 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.949577 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xr4k"] Dec 04 20:34:56 crc kubenswrapper[4733]: I1204 20:34:56.967185 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4k" event={"ID":"0a5b223e-fe0f-4593-9a83-e81b54be6a63","Type":"ContainerStarted","Data":"63565e8b9c0452cb4768ae080e4ba2e134c12a9e550931d13a2b8a854eaa19c5"} Dec 04 20:34:57 crc kubenswrapper[4733]: I1204 20:34:57.993085 4733 generic.go:334] "Generic (PLEG): container finished" podID="0a5b223e-fe0f-4593-9a83-e81b54be6a63" containerID="976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e" exitCode=0 Dec 04 20:34:57 crc kubenswrapper[4733]: I1204 20:34:57.993154 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4k" event={"ID":"0a5b223e-fe0f-4593-9a83-e81b54be6a63","Type":"ContainerDied","Data":"976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e"} Dec 04 20:35:00 crc kubenswrapper[4733]: I1204 20:35:00.017097 4733 generic.go:334] "Generic (PLEG): container finished" podID="0a5b223e-fe0f-4593-9a83-e81b54be6a63" containerID="bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274" exitCode=0 Dec 04 20:35:00 crc kubenswrapper[4733]: I1204 20:35:00.017705 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4k" event={"ID":"0a5b223e-fe0f-4593-9a83-e81b54be6a63","Type":"ContainerDied","Data":"bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274"} Dec 04 20:35:01 crc kubenswrapper[4733]: I1204 20:35:01.028260 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4k" event={"ID":"0a5b223e-fe0f-4593-9a83-e81b54be6a63","Type":"ContainerStarted","Data":"8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733"} Dec 04 20:35:01 crc kubenswrapper[4733]: I1204 20:35:01.055125 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6xr4k" podStartSLOduration=2.511534642 podStartE2EDuration="5.055108383s" podCreationTimestamp="2025-12-04 20:34:56 +0000 UTC" firstStartedPulling="2025-12-04 20:34:58.002686528 +0000 UTC m=+10559.958047584" lastFinishedPulling="2025-12-04 20:35:00.546260269 +0000 UTC m=+10562.501621325" observedRunningTime="2025-12-04 20:35:01.054278381 +0000 UTC m=+10563.009639427" watchObservedRunningTime="2025-12-04 20:35:01.055108383 +0000 UTC m=+10563.010469429" Dec 04 20:35:06 crc kubenswrapper[4733]: I1204 20:35:06.428310 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:35:06 crc kubenswrapper[4733]: I1204 20:35:06.429233 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:35:06 crc kubenswrapper[4733]: I1204 20:35:06.491321 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:35:07 crc kubenswrapper[4733]: I1204 20:35:07.136486 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:35:07 crc kubenswrapper[4733]: I1204 20:35:07.189101 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xr4k"] Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.110754 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6xr4k" podUID="0a5b223e-fe0f-4593-9a83-e81b54be6a63" containerName="registry-server" containerID="cri-o://8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733" gracePeriod=2 Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.695219 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.745600 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-utilities\") pod \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.745708 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8f44\" (UniqueName: \"kubernetes.io/projected/0a5b223e-fe0f-4593-9a83-e81b54be6a63-kube-api-access-n8f44\") pod \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.746036 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-catalog-content\") pod \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\" (UID: \"0a5b223e-fe0f-4593-9a83-e81b54be6a63\") " Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.747167 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-utilities" (OuterVolumeSpecName: "utilities") pod "0a5b223e-fe0f-4593-9a83-e81b54be6a63" (UID: "0a5b223e-fe0f-4593-9a83-e81b54be6a63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.754149 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a5b223e-fe0f-4593-9a83-e81b54be6a63-kube-api-access-n8f44" (OuterVolumeSpecName: "kube-api-access-n8f44") pod "0a5b223e-fe0f-4593-9a83-e81b54be6a63" (UID: "0a5b223e-fe0f-4593-9a83-e81b54be6a63"). InnerVolumeSpecName "kube-api-access-n8f44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.849569 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.849609 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8f44\" (UniqueName: \"kubernetes.io/projected/0a5b223e-fe0f-4593-9a83-e81b54be6a63-kube-api-access-n8f44\") on node \"crc\" DevicePath \"\"" Dec 04 20:35:09 crc kubenswrapper[4733]: I1204 20:35:09.974760 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a5b223e-fe0f-4593-9a83-e81b54be6a63" (UID: "0a5b223e-fe0f-4593-9a83-e81b54be6a63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.053667 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5b223e-fe0f-4593-9a83-e81b54be6a63-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.124390 4733 generic.go:334] "Generic (PLEG): container finished" podID="0a5b223e-fe0f-4593-9a83-e81b54be6a63" containerID="8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733" exitCode=0 Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.124441 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4k" event={"ID":"0a5b223e-fe0f-4593-9a83-e81b54be6a63","Type":"ContainerDied","Data":"8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733"} Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.124466 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xr4k" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.124489 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xr4k" event={"ID":"0a5b223e-fe0f-4593-9a83-e81b54be6a63","Type":"ContainerDied","Data":"63565e8b9c0452cb4768ae080e4ba2e134c12a9e550931d13a2b8a854eaa19c5"} Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.124512 4733 scope.go:117] "RemoveContainer" containerID="8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.147366 4733 scope.go:117] "RemoveContainer" containerID="bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.169693 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xr4k"] Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.180238 4733 scope.go:117] "RemoveContainer" containerID="976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.185153 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6xr4k"] Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.230417 4733 scope.go:117] "RemoveContainer" containerID="8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733" Dec 04 20:35:10 crc kubenswrapper[4733]: E1204 20:35:10.231568 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733\": container with ID starting with 8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733 not found: ID does not exist" containerID="8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.231619 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733"} err="failed to get container status \"8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733\": rpc error: code = NotFound desc = could not find container \"8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733\": container with ID starting with 8daf32c7842d954178e2212432569f792aca9bda8125fd61bb7bd6328a2d4733 not found: ID does not exist" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.231646 4733 scope.go:117] "RemoveContainer" containerID="bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274" Dec 04 20:35:10 crc kubenswrapper[4733]: E1204 20:35:10.232316 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274\": container with ID starting with bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274 not found: ID does not exist" containerID="bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.232391 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274"} err="failed to get container status \"bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274\": rpc error: code = NotFound desc = could not find container \"bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274\": container with ID starting with bcc941229e71c096f1b1f5e64e71e60450242def2f7e88ee189393db3b2a1274 not found: ID does not exist" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.232424 4733 scope.go:117] "RemoveContainer" containerID="976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e" Dec 04 20:35:10 crc kubenswrapper[4733]: E1204 20:35:10.232897 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e\": container with ID starting with 976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e not found: ID does not exist" containerID="976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.232958 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e"} err="failed to get container status \"976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e\": rpc error: code = NotFound desc = could not find container \"976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e\": container with ID starting with 976ce846672ad1dfb7dc7d31d45b8dd229ce0f958c34a17d0595d6fa6bd7ae1e not found: ID does not exist" Dec 04 20:35:10 crc kubenswrapper[4733]: I1204 20:35:10.347785 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a5b223e-fe0f-4593-9a83-e81b54be6a63" path="/var/lib/kubelet/pods/0a5b223e-fe0f-4593-9a83-e81b54be6a63/volumes" Dec 04 20:35:15 crc kubenswrapper[4733]: I1204 20:35:15.362476 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:35:15 crc kubenswrapper[4733]: I1204 20:35:15.362858 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:35:45 crc kubenswrapper[4733]: I1204 20:35:45.361919 4733 patch_prober.go:28] interesting pod/machine-config-daemon-w7n9m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 20:35:45 crc kubenswrapper[4733]: I1204 20:35:45.362618 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 20:35:45 crc kubenswrapper[4733]: I1204 20:35:45.362681 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" Dec 04 20:35:45 crc kubenswrapper[4733]: I1204 20:35:45.363989 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f"} pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 20:35:45 crc kubenswrapper[4733]: I1204 20:35:45.364071 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerName="machine-config-daemon" containerID="cri-o://e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" gracePeriod=600 Dec 04 20:35:46 crc kubenswrapper[4733]: I1204 20:35:46.536302 4733 generic.go:334] "Generic (PLEG): container finished" podID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" exitCode=0 Dec 04 20:35:46 crc kubenswrapper[4733]: I1204 20:35:46.536385 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" event={"ID":"5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9","Type":"ContainerDied","Data":"e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f"} Dec 04 20:35:46 crc kubenswrapper[4733]: I1204 20:35:46.536732 4733 scope.go:117] "RemoveContainer" containerID="fb18a7d2636425cdb65bf7d0042b43bec6a6390071894b7c5044793db67a3f33" Dec 04 20:35:46 crc kubenswrapper[4733]: E1204 20:35:46.806405 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:35:47 crc kubenswrapper[4733]: I1204 20:35:47.553510 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:35:47 crc kubenswrapper[4733]: E1204 20:35:47.554914 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:35:59 crc kubenswrapper[4733]: I1204 20:35:59.335778 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:35:59 crc kubenswrapper[4733]: E1204 20:35:59.336926 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:36:13 crc kubenswrapper[4733]: I1204 20:36:13.336575 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:36:13 crc kubenswrapper[4733]: E1204 20:36:13.337846 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:36:24 crc kubenswrapper[4733]: I1204 20:36:24.335838 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:36:24 crc kubenswrapper[4733]: E1204 20:36:24.339731 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:36:36 crc kubenswrapper[4733]: I1204 20:36:36.336018 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:36:36 crc kubenswrapper[4733]: E1204 20:36:36.336952 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:36:51 crc kubenswrapper[4733]: I1204 20:36:51.336342 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:36:51 crc kubenswrapper[4733]: E1204 20:36:51.337174 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:37:05 crc kubenswrapper[4733]: I1204 20:37:05.335788 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:37:05 crc kubenswrapper[4733]: E1204 20:37:05.336566 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:37:17 crc kubenswrapper[4733]: I1204 20:37:17.336047 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:37:17 crc kubenswrapper[4733]: E1204 20:37:17.337136 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:37:29 crc kubenswrapper[4733]: I1204 20:37:29.335921 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:37:29 crc kubenswrapper[4733]: E1204 20:37:29.336918 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:37:40 crc kubenswrapper[4733]: I1204 20:37:40.336124 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:37:40 crc kubenswrapper[4733]: E1204 20:37:40.337089 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:37:54 crc kubenswrapper[4733]: I1204 20:37:54.335357 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:37:54 crc kubenswrapper[4733]: E1204 20:37:54.336172 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:38:07 crc kubenswrapper[4733]: I1204 20:38:07.335373 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:38:07 crc kubenswrapper[4733]: E1204 20:38:07.336188 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" Dec 04 20:38:22 crc kubenswrapper[4733]: I1204 20:38:22.336672 4733 scope.go:117] "RemoveContainer" containerID="e0f13fecb99468b45affcf3a3b4f21e6a943b6670fbbac9f766d68a7e08f5c1f" Dec 04 20:38:22 crc kubenswrapper[4733]: E1204 20:38:22.337324 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w7n9m_openshift-machine-config-operator(5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9)\"" pod="openshift-machine-config-operator/machine-config-daemon-w7n9m" podUID="5a6332b0-c20e-4bca-91c5-2f2dcedcbbc9" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114370515024447 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114370515017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114343116016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114343116015454 5ustar corecore